2025-09-07T07:35:07.6535071Z Current runner version: '2.328.0' 2025-09-07T07:35:07.6539980Z Runner name: 'i-0d9447236daff7b96' 2025-09-07T07:35:07.6540771Z Runner group name: 'default' 2025-09-07T07:35:07.6541624Z Machine name: 'ip-10-0-57-198' 2025-09-07T07:35:07.6543782Z ##[group]GITHUB_TOKEN Permissions 2025-09-07T07:35:07.6545678Z Contents: read 2025-09-07T07:35:07.6546300Z Metadata: read 2025-09-07T07:35:07.6546708Z ##[endgroup] 2025-09-07T07:35:07.6548391Z Secret source: Actions 2025-09-07T07:35:07.6548928Z Prepare workflow directory 2025-09-07T07:35:07.6933134Z Prepare all required actions 2025-09-07T07:35:07.6962390Z Getting action download info 2025-09-07T07:35:08.4087844Z Download action repository 'pytorch/test-infra@main' (SHA:548a4bc624d43a01cdf165a63b041f0ae014ddbd) 2025-09-07T07:35:09.4641930Z Download action repository 'pytorch/pytorch@main' (SHA:93fb23d6fae7c4e82c4239a1033e522088742634) 2025-09-07T07:35:23.0925962Z Download action repository 'actions/setup-python@a26af69be951a213d495a4c3e4e4022e16d87065' (SHA:a26af69be951a213d495a4c3e4e4022e16d87065) 2025-09-07T07:35:23.3385046Z Download action repository 'aws-actions/configure-aws-credentials@ececac1a45f3b08a01d2dd070d28d111c5fe6722' (SHA:ececac1a45f3b08a01d2dd070d28d111c5fe6722) 2025-09-07T07:35:23.6070316Z Download action repository 'aws-actions/amazon-ecr-login@062b18b96a7aff071d4dc91bc00c4c1a7945b076' (SHA:062b18b96a7aff071d4dc91bc00c4c1a7945b076) 2025-09-07T07:35:23.7396277Z Download action repository 'seemethere/upload-artifact-s3@baba72d0712b404f646cebe0730933554ebce96a' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-09-07T07:35:24.0198874Z Getting action download info 2025-09-07T07:35:24.1125261Z Download action repository 'actions/checkout@v4' (SHA:08eba0b27e820071cde6df949e0beb9ba4906955) 2025-09-07T07:35:24.3658374Z Getting action download info 2025-09-07T07:35:24.4720189Z Download action repository 'nick-fields/retry@v3.0.0' (SHA:7152eba30c6575329ac0576536151aca5a72780e) 2025-09-07T07:35:24.6506543Z Getting action download info 2025-09-07T07:35:24.7806574Z Download action repository 'nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482' (SHA:3e91a01664abd3c5cd539100d10d33b9c5b68482) 2025-09-07T07:35:25.0517349Z Getting action download info 2025-09-07T07:35:25.1716675Z Uses: pytorch/pytorch/.github/workflows/_linux-test.yml@refs/heads/main (93fb23d6fae7c4e82c4239a1033e522088742634) 2025-09-07T07:35:25.1719670Z ##[group] Inputs 2025-09-07T07:35:25.1719951Z build-environment: linux-jammy-py3.9-gcc11-build 2025-09-07T07:35:25.1722461Z test-matrix: {"include": [{"config": "inductor_huggingface_perf_cpu_x86_zen", "shard": 1, "num_shards": 3, "runner": "linux.24xlarge.amd"}, {"config": "inductor_huggingface_perf_cpu_x86_zen", "shard": 2, "num_shards": 3, "runner": "linux.24xlarge.amd"}, {"config": "inductor_huggingface_perf_cpu_x86_zen", "shard": 3, "num_shards": 3, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 1, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 2, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 3, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 4, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 5, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 1, "num_shards": 4, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 2, "num_shards": 4, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 3, "num_shards": 4, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 4, "num_shards": 4, "runner": "linux.24xlarge.amd"}]} 2025-09-07T07:35:25.1725756Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:35:25.1726455Z sync-tag: 2025-09-07T07:35:25.1727122Z timeout-minutes: 720 2025-09-07T07:35:25.1727296Z use-gha: 2025-09-07T07:35:25.1727691Z dashboard-tag: training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true 2025-09-07T07:35:25.1728124Z s3-bucket: gha-artifacts 2025-09-07T07:35:25.1728300Z aws-role-to-assume: 2025-09-07T07:35:25.1728688Z disable-monitor: false 2025-09-07T07:35:25.1728891Z monitor-log-interval: 15 2025-09-07T07:35:25.1729097Z monitor-data-collect-interval: 4 2025-09-07T07:35:25.1729300Z ##[endgroup] 2025-09-07T07:35:25.1729628Z Complete job name: inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd) 2025-09-07T07:35:25.2811225Z A job started hook has been configured by the self-hosted runner administrator 2025-09-07T07:35:25.2888576Z ##[group]Run '/home/ec2-user/runner-scripts/before_job.sh' 2025-09-07T07:35:25.2897118Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:35:25.2897578Z ##[endgroup] 2025-09-07T07:35:26.4664054Z Runner Type: linux.24xlarge.amd 2025-09-07T07:35:26.4664399Z Instance Type: m7a.24xlarge 2025-09-07T07:35:26.4664708Z AMI Name: unknown 2025-09-07T07:35:26.4704034Z AMI ID: ami-05ffe3c48a9991133 2025-09-07T07:35:30.7350398Z ##[group]Run pytorch/test-infra/.github/actions/setup-ssh@main 2025-09-07T07:35:30.7350733Z with: 2025-09-07T07:35:30.7351190Z github-secret: *** 2025-09-07T07:35:30.7351638Z instructions: All testing is done inside the container, to start an interactive session run: docker exec -it $(docker container ps --format '{{.ID}}') bash 2025-09-07T07:35:30.7352101Z activate-with-label: false 2025-09-07T07:35:30.7352290Z label: with-ssh 2025-09-07T07:35:30.7352460Z remove-existing-keys: true 2025-09-07T07:35:30.7352631Z fail-silently: true 2025-09-07T07:35:30.7352789Z env: 2025-09-07T07:35:30.7352935Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:35:30.7353113Z ##[endgroup] 2025-09-07T07:35:30.8440392Z Please see https://github.com/pytorch/pytorch/wiki/Debugging-using-with-ssh-for-Github-Actions for more info. 2025-09-07T07:35:30.8441064Z Not on pull request and ciflow reference could not be extracted, skipping adding ssh keys 2025-09-07T07:35:30.8609378Z ##[group]Run pytorch/pytorch/.github/actions/checkout-pytorch@main 2025-09-07T07:35:30.8609678Z with: 2025-09-07T07:35:30.8609815Z no-sudo: true 2025-09-07T07:35:30.8609985Z submodules: recursive 2025-09-07T07:35:30.8610149Z fetch-depth: 0 2025-09-07T07:35:30.8610356Z env: 2025-09-07T07:35:30.8610490Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:35:30.8610684Z ##[endgroup] 2025-09-07T07:35:30.8664336Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:35:30.8664945Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:35:30.8676551Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:35:30.8676809Z env: 2025-09-07T07:35:30.8676983Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:35:30.8677197Z ##[endgroup] 2025-09-07T07:35:30.8761805Z ##[group]Run # Use all available CPUs for fetching 2025-09-07T07:35:30.8762120Z # Use all available CPUs for fetching 2025-09-07T07:35:30.8762356Z cd "${GITHUB_WORKSPACE}" 2025-09-07T07:35:30.8762582Z git config --global fetch.parallel 0 2025-09-07T07:35:30.8762832Z git config --global submodule.fetchJobs 0 2025-09-07T07:35:30.8763049Z  2025-09-07T07:35:30.8763282Z # Clean workspace. The default checkout action should also do this, but 2025-09-07T07:35:30.8763569Z # do it here as well just in case 2025-09-07T07:35:30.8763792Z if [[ -d .git ]]; then 2025-09-07T07:35:30.8764003Z  if [ -z "${NO_SUDO}" ]; then 2025-09-07T07:35:30.8764204Z  sudo git clean -ffdx 2025-09-07T07:35:30.8764382Z  else 2025-09-07T07:35:30.8764766Z  git clean -ffdx 2025-09-07T07:35:30.8764943Z  fi 2025-09-07T07:35:30.8765095Z fi 2025-09-07T07:35:30.8772204Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:35:30.8772449Z env: 2025-09-07T07:35:30.8772682Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:35:30.8772880Z NO_SUDO: true 2025-09-07T07:35:30.8773053Z ##[endgroup] 2025-09-07T07:35:30.8880652Z ##[group]Run actions/checkout@v4 2025-09-07T07:35:30.8880863Z with: 2025-09-07T07:35:30.8881040Z ref: 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:35:30.8881254Z fetch-depth: 0 2025-09-07T07:35:30.8881423Z submodules: recursive 2025-09-07T07:35:30.8881597Z show-progress: false 2025-09-07T07:35:30.8881774Z repository: pytorch/pytorch 2025-09-07T07:35:30.8882032Z token: *** 2025-09-07T07:35:30.8882183Z ssh-strict: true 2025-09-07T07:35:30.8882343Z ssh-user: git 2025-09-07T07:35:30.8882508Z persist-credentials: true 2025-09-07T07:35:30.8882680Z clean: true 2025-09-07T07:35:30.8882865Z sparse-checkout-cone-mode: true 2025-09-07T07:35:30.8883061Z fetch-tags: false 2025-09-07T07:35:30.8883215Z lfs: false 2025-09-07T07:35:30.8883356Z set-safe-directory: true 2025-09-07T07:35:30.8883535Z env: 2025-09-07T07:35:30.8883676Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:35:30.8883839Z ##[endgroup] 2025-09-07T07:35:30.9762172Z Syncing repository: pytorch/pytorch 2025-09-07T07:35:30.9763185Z ##[group]Getting Git version info 2025-09-07T07:35:30.9763512Z Working directory is '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-09-07T07:35:30.9763941Z [command]/usr/bin/git version 2025-09-07T07:35:30.9986542Z git version 2.47.1 2025-09-07T07:35:31.0006753Z ##[endgroup] 2025-09-07T07:35:31.0015840Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/ee3f3f1a-b755-4d9e-b4d6-8a7a394b087b/.gitconfig' 2025-09-07T07:35:31.0032824Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/ee3f3f1a-b755-4d9e-b4d6-8a7a394b087b' before making global git config changes 2025-09-07T07:35:31.0033443Z Adding repository directory to the temporary git global config as a safe directory 2025-09-07T07:35:31.0037258Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-09-07T07:35:31.0095017Z Deleting the contents of '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-09-07T07:35:31.0097986Z ##[group]Initializing the repository 2025-09-07T07:35:31.0101318Z [command]/usr/bin/git init /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-09-07T07:35:31.0167272Z hint: Using 'master' as the name for the initial branch. This default branch name 2025-09-07T07:35:31.0167651Z hint: is subject to change. To configure the initial branch name to use in all 2025-09-07T07:35:31.0168481Z hint: of your new repositories, which will suppress this warning, call: 2025-09-07T07:35:31.0168745Z hint: 2025-09-07T07:35:31.0168977Z hint: git config --global init.defaultBranch 2025-09-07T07:35:31.0169213Z hint: 2025-09-07T07:35:31.0169426Z hint: Names commonly chosen instead of 'master' are 'main', 'trunk' and 2025-09-07T07:35:31.0169780Z hint: 'development'. The just-created branch can be renamed via this command: 2025-09-07T07:35:31.0170044Z hint: 2025-09-07T07:35:31.0170201Z hint: git branch -m 2025-09-07T07:35:31.0187419Z Initialized empty Git repository in /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/ 2025-09-07T07:35:31.0196891Z [command]/usr/bin/git remote add origin https://github.com/pytorch/pytorch 2025-09-07T07:35:31.0240049Z ##[endgroup] 2025-09-07T07:35:31.0240411Z ##[group]Disabling automatic garbage collection 2025-09-07T07:35:31.0242827Z [command]/usr/bin/git config --local gc.auto 0 2025-09-07T07:35:31.0269528Z ##[endgroup] 2025-09-07T07:35:31.0269782Z ##[group]Setting up auth 2025-09-07T07:35:31.0274542Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-09-07T07:35:31.0304368Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-09-07T07:35:31.0711245Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-09-07T07:35:31.0737801Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-09-07T07:35:31.1086759Z [command]/usr/bin/git config --local http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-09-07T07:35:31.1130887Z ##[endgroup] 2025-09-07T07:35:31.1131276Z ##[group]Fetching the repository 2025-09-07T07:35:31.1136599Z [command]/usr/bin/git -c protocol.version=2 fetch --prune --no-recurse-submodules origin +refs/heads/*:refs/remotes/origin/* +refs/tags/*:refs/tags/* 2025-09-07T07:36:03.6298253Z From https://github.com/pytorch/pytorch 2025-09-07T07:36:03.6298615Z * [new branch] 160583 -> origin/160583 2025-09-07T07:36:03.6299099Z * [new branch] 2.6.0.dev20241004+ -> origin/2.6.0.dev20241004+ 2025-09-07T07:36:03.6299439Z * [new branch] 5addvllmbuild -> origin/5addvllmbuild 2025-09-07T07:36:03.6299824Z * [new branch] AaronWang04_addmmfusion_perftest -> origin/AaronWang04_addmmfusion_perftest 2025-09-07T07:36:03.6300904Z * [new branch] HDCharles-2.6.0-release-notes -> origin/HDCharles-2.6.0-release-notes 2025-09-07T07:36:03.6301927Z * [new branch] ISSUE-154849 -> origin/ISSUE-154849 2025-09-07T07:36:03.6304498Z * [new branch] JackCaoG/dynamo_make_fx_non_core_aten_ops -> origin/JackCaoG/dynamo_make_fx_non_core_aten_ops 2025-09-07T07:36:03.6306255Z * [new branch] NicoshevSVE128 -> origin/NicoshevSVE128 2025-09-07T07:36:03.6307597Z * [new branch] PR-AOTInductorNoneBug -> origin/PR-AOTInductorNoneBug 2025-09-07T07:36:03.6308863Z * [new branch] PR-AOTInductorNoneBugFix -> origin/PR-AOTInductorNoneBugFix 2025-09-07T07:36:03.6310935Z * [new branch] PR-FixConfigsIssue -> origin/PR-FixConfigsIssue 2025-09-07T07:36:03.6311300Z * [new branch] PR-NoneBugFix-viable -> origin/PR-NoneBugFix-viable 2025-09-07T07:36:03.6312555Z * [new branch] PR-ResetToZero -> origin/PR-ResetToZero 2025-09-07T07:36:03.6313866Z * [new branch] Update-Flash-Packaging -> origin/Update-Flash-Packaging 2025-09-07T07:36:03.6315084Z * [new branch] VLA_exp -> origin/VLA_exp 2025-09-07T07:36:03.6316578Z * [new branch] actually-run-mps-aot-inductor -> origin/actually-run-mps-aot-inductor 2025-09-07T07:36:03.6317988Z * [new branch] add-missing-args-normalization -> origin/add-missing-args-normalization 2025-09-07T07:36:03.6319268Z * [new branch] add-user-guide-structure -> origin/add-user-guide-structure 2025-09-07T07:36:03.6320738Z * [new branch] add-vllm-nightly-build -> origin/add-vllm-nightly-build 2025-09-07T07:36:03.6321902Z * [new branch] add_compile_benchmarking -> origin/add_compile_benchmarking 2025-09-07T07:36:03.6323161Z * [new branch] addmm-heuristic -> origin/addmm-heuristic 2025-09-07T07:36:03.6324447Z * [new branch] addsimde -> origin/addsimde 2025-09-07T07:36:03.6325738Z * [new branch] addvllmtest -> origin/addvllmtest 2025-09-07T07:36:03.6327680Z * [new branch] adi/acl_upgrade -> origin/adi/acl_upgrade 2025-09-07T07:36:03.6328914Z * [new branch] adi/test -> origin/adi/test 2025-09-07T07:36:03.6330221Z * [new branch] adi/test_bgemm -> origin/adi/test_bgemm 2025-09-07T07:36:03.6331607Z * [new branch] adi/test_fusions -> origin/adi/test_fusions 2025-09-07T07:36:03.6335555Z * [new branch] adi/test_onednn_v3.9 -> origin/adi/test_onednn_v3.9 2025-09-07T07:36:03.6336212Z * [new branch] adi/test_presve_change -> origin/adi/test_presve_change 2025-09-07T07:36:03.6336548Z * [new branch] adi/test_timm -> origin/adi/test_timm 2025-09-07T07:36:03.6337668Z * [new branch] adi/testpresve_change -> origin/adi/testpresve_change 2025-09-07T07:36:03.6339674Z * [new branch] aditew01/test/vec_bf16 -> origin/aditew01/test/vec_bf16 2025-09-07T07:36:03.6340972Z * [new branch] ah-globalfeedback-hook -> origin/ah-globalfeedback-hook 2025-09-07T07:36:03.6342253Z * [new branch] alt-disable -> origin/alt-disable 2025-09-07T07:36:03.6344177Z * [new branch] angelayi/aoti_additional_files -> origin/angelayi/aoti_additional_files 2025-09-07T07:36:03.6345409Z * [new branch] angelayi/aoti_inductor_fx -> origin/angelayi/aoti_inductor_fx 2025-09-07T07:36:03.6346559Z * [new branch] angelayi/benchmark -> origin/angelayi/benchmark 2025-09-07T07:36:03.6347921Z * [new branch] angelayi/benchmark2 -> origin/angelayi/benchmark2 2025-09-07T07:36:03.6349164Z * [new branch] angelayi/change_pytree_serialization -> origin/angelayi/change_pytree_serialization 2025-09-07T07:36:03.6350328Z * [new branch] angelayi/cpp_loader -> origin/angelayi/cpp_loader 2025-09-07T07:36:03.6351999Z * [new branch] angelayi/custom_op_subgraph -> origin/angelayi/custom_op_subgraph 2025-09-07T07:36:03.6353566Z * [new branch] angelayi/customop -> origin/angelayi/customop 2025-09-07T07:36:03.6355157Z * [new branch] angelayi/fake_cache_empty -> origin/angelayi/fake_cache_empty 2025-09-07T07:36:03.6356468Z * [new branch] angelayi/is_symbolic_tracing -> origin/angelayi/is_symbolic_tracing 2025-09-07T07:36:03.6357673Z * [new branch] angelayi/item -> origin/angelayi/item 2025-09-07T07:36:03.6359075Z * [new branch] angelayi/no_so_weight -> origin/angelayi/no_so_weight 2025-09-07T07:36:03.6360171Z * [new branch] angelayi/opoverload -> origin/angelayi/opoverload 2025-09-07T07:36:03.6361445Z * [new branch] angelayi/pattern -> origin/angelayi/pattern 2025-09-07T07:36:03.6362747Z * [new branch] angelayi/pytree -> origin/angelayi/pytree 2025-09-07T07:36:03.6364022Z * [new branch] angelayi/scan_layers -> origin/angelayi/scan_layers 2025-09-07T07:36:03.6365274Z * [new branch] angelayi/symint_input -> origin/angelayi/symint_input 2025-09-07T07:36:03.6366547Z * [new branch] angelayi/test_cpp -> origin/angelayi/test_cpp 2025-09-07T07:36:03.6367786Z * [new branch] angelayi/torch_size -> origin/angelayi/torch_size 2025-09-07T07:36:03.6369066Z * [new branch] aoti-cuda-alloc -> origin/aoti-cuda-alloc 2025-09-07T07:36:03.6370341Z * [new branch] aoti_target_windows -> origin/aoti_target_windows 2025-09-07T07:36:03.6371590Z * [new branch] aoti_weight_sharing -> origin/aoti_weight_sharing 2025-09-07T07:36:03.6373001Z * [new branch] atalman-inductor-perf-cu124 -> origin/atalman-inductor-perf-cu124 2025-09-07T07:36:03.6374225Z * [new branch] atalman-inductor-perf-cu124.1 -> origin/atalman-inductor-perf-cu124.1 2025-09-07T07:36:03.6375520Z * [new branch] atalman-patch-1 -> origin/atalman-patch-1 2025-09-07T07:36:03.6376846Z * [new branch] atalman-patch-3 -> origin/atalman-patch-3 2025-09-07T07:36:03.6378198Z * [new branch] atalman-patch-4 -> origin/atalman-patch-4 2025-09-07T07:36:03.6379523Z * [new branch] atalman-patch-5 -> origin/atalman-patch-5 2025-09-07T07:36:03.6380798Z * [new branch] atalman-patch-6 -> origin/atalman-patch-6 2025-09-07T07:36:03.6382084Z * [new branch] atalman_inductor_2.3.0 -> origin/atalman_inductor_2.3.0 2025-09-07T07:36:03.6383304Z * [new branch] atalman_inductor_2.3.1 -> origin/atalman_inductor_2.3.1 2025-09-07T07:36:03.6384517Z * [new branch] atalman_inductor_2.4.0 -> origin/atalman_inductor_2.4.0 2025-09-07T07:36:03.6385944Z * [new branch] atalman_inductor_2.4.x -> origin/atalman_inductor_2.4.x 2025-09-07T07:36:03.6387331Z * [new branch] autoupdate-transformers-pin-via-pr -> origin/autoupdate-transformers-pin-via-pr 2025-09-07T07:36:03.6388996Z * [new branch] bahuang/dtensor_demo -> origin/bahuang/dtensor_demo 2025-09-07T07:36:03.6390227Z * [new branch] bahuang/test -> origin/bahuang/test 2025-09-07T07:36:03.6392267Z * [new branch] base/1.5 -> origin/base/1.5 2025-09-07T07:36:03.6393585Z * [new branch] batching_sdpa_efficient_attention -> origin/batching_sdpa_efficient_attention 2025-09-07T07:36:03.6394792Z * [new branch] bc-lint-config -> origin/bc-lint-config 2025-09-07T07:36:03.6396047Z * [new branch] bc-lint-test-new-config -> origin/bc-lint-test-new-config 2025-09-07T07:36:03.6397373Z * [new branch] benchmark-updates -> origin/benchmark-updates 2025-09-07T07:36:03.6398690Z * [new branch] benchmarker_compat_with_do_bench -> origin/benchmarker_compat_with_do_bench 2025-09-07T07:36:03.6399934Z * [new branch] benchmarking-script -> origin/benchmarking-script 2025-09-07T07:36:03.6401768Z * [new branch] bertmaher/pinbump26 -> origin/bertmaher/pinbump26 2025-09-07T07:36:03.6403802Z * [new branch] bertrand/cutlass -> origin/bertrand/cutlass 2025-09-07T07:36:03.6405593Z * [new branch] bf/cg-custom-wrapper -> origin/bf/cg-custom-wrapper 2025-09-07T07:36:03.6406892Z * [new branch] bf/cg-or-error -> origin/bf/cg-or-error 2025-09-07T07:36:03.6407932Z * [new branch] bf/cg-remove-check -> origin/bf/cg-remove-check 2025-09-07T07:36:03.6409162Z * [new branch] bf/cg-skip-1-kernel -> origin/bf/cg-skip-1-kernel 2025-09-07T07:36:03.6410378Z * [new branch] bf/cudagraph -> origin/bf/cudagraph 2025-09-07T07:36:03.6411870Z * [new branch] bf/cudagraph-disable-input-mutation -> origin/bf/cudagraph-disable-input-mutation 2025-09-07T07:36:03.6413656Z * [new branch] bf/cudagraph-enable-input-mutation-support-benchmark -> origin/bf/cudagraph-enable-input-mutation-support-benchmark 2025-09-07T07:36:03.6414705Z * [new branch] bf/cudagraph-partition -> origin/bf/cudagraph-partition 2025-09-07T07:36:03.6416077Z * [new branch] bf/default-recompile-reason -> origin/bf/default-recompile-reason 2025-09-07T07:36:03.6417312Z * [new branch] bf/donated-buffer-bench -> origin/bf/donated-buffer-bench 2025-09-07T07:36:03.6418662Z * [new branch] bf/exp -> origin/bf/exp 2025-09-07T07:36:03.6419939Z * [new branch] bf/pa-non-divisible -> origin/bf/pa-non-divisible 2025-09-07T07:36:03.6421330Z * [new branch] bf/partition-move-cpu -> origin/bf/partition-move-cpu 2025-09-07T07:36:03.6422641Z * [new branch] bf/partition-turn-on -> origin/bf/partition-turn-on 2025-09-07T07:36:03.6423848Z * [new branch] bf/remove-check-55b0c39d -> origin/bf/remove-check-55b0c39d 2025-09-07T07:36:03.6424989Z * [new branch] bf/rope -> origin/bf/rope 2025-09-07T07:36:03.6426294Z * [new branch] bisect_perf_hf_T5_3acc6eac492 -> origin/bisect_perf_hf_T5_3acc6eac492 2025-09-07T07:36:03.6427548Z * [new branch] bisect_perf_hf_T5_3fcf66f61fb -> origin/bisect_perf_hf_T5_3fcf66f61fb 2025-09-07T07:36:03.6428703Z * [new branch] bisect_perf_hf_T5_4009d154129 -> origin/bisect_perf_hf_T5_4009d154129 2025-09-07T07:36:03.6429859Z * [new branch] bisect_perf_hf_T5_40d0740e73d -> origin/bisect_perf_hf_T5_40d0740e73d 2025-09-07T07:36:03.6431029Z * [new branch] bisect_perf_hf_T5_5268754e -> origin/bisect_perf_hf_T5_5268754e 2025-09-07T07:36:03.6432231Z * [new branch] bisect_perf_hf_T5_7d89a8d385c -> origin/bisect_perf_hf_T5_7d89a8d385c 2025-09-07T07:36:03.6433377Z * [new branch] bisect_perf_hf_T5_b7a25c1ee7c -> origin/bisect_perf_hf_T5_b7a25c1ee7c 2025-09-07T07:36:03.6434673Z * [new branch] bisect_perf_hf_T5_c25b201583f -> origin/bisect_perf_hf_T5_c25b201583f 2025-09-07T07:36:03.6435988Z * [new branch] bisect_perf_hf_T5_c93e57efac0 -> origin/bisect_perf_hf_T5_c93e57efac0 2025-09-07T07:36:03.6437333Z * [new branch] bisect_perf_hf_T5_ca9813ea149 -> origin/bisect_perf_hf_T5_ca9813ea149 2025-09-07T07:36:03.6438665Z * [new branch] bisect_perf_hf_T5_d65f194a -> origin/bisect_perf_hf_T5_d65f194a 2025-09-07T07:36:03.6439883Z * [new branch] bisect_perf_hf_T5_da94ab0b -> origin/bisect_perf_hf_T5_da94ab0b 2025-09-07T07:36:03.6441089Z * [new branch] bisect_perf_hf_T5_da94ab0b_new -> origin/bisect_perf_hf_T5_da94ab0b_new 2025-09-07T07:36:03.6442289Z * [new branch] bisect_perf_hf_T5_db4e8a1d8a8 -> origin/bisect_perf_hf_T5_db4e8a1d8a8 2025-09-07T07:36:03.6443479Z * [new branch] bisect_perf_hf_T5_e0d97e936a2 -> origin/bisect_perf_hf_T5_e0d97e936a2 2025-09-07T07:36:03.6444736Z * [new branch] bisect_perf_hf_T5_f23621ec563 -> origin/bisect_perf_hf_T5_f23621ec563 2025-09-07T07:36:03.6446628Z * [new branch] bowbao/bench_updates_stage -> origin/bowbao/bench_updates_stage 2025-09-07T07:36:03.6447867Z * [new branch] bowbao/dort_rewriter -> origin/bowbao/dort_rewriter 2025-09-07T07:36:03.6449070Z * [new branch] bowbao/wip_prs -> origin/bowbao/wip_prs 2025-09-07T07:36:03.6450732Z * [new branch] brister/break_tensorbox -> origin/brister/break_tensorbox 2025-09-07T07:36:03.6451899Z * [new branch] brister/custom_fx_backend -> origin/brister/custom_fx_backend 2025-09-07T07:36:03.6453174Z * [new branch] brister/fx_custom_triton -> origin/brister/fx_custom_triton 2025-09-07T07:36:03.6454319Z * [new branch] brister/tensor_box_output -> origin/brister/tensor_box_output 2025-09-07T07:36:03.6455530Z * [new branch] brister/tiled_reduction_no_numel_check -> origin/brister/tiled_reduction_no_numel_check 2025-09-07T07:36:03.6456781Z * [new branch] c57382a49 -> origin/c57382a49 2025-09-07T07:36:03.6458062Z * [new branch] ca_0431d47eaa -> origin/ca_0431d47eaa 2025-09-07T07:36:03.6459272Z * [new branch] ca_fix_0431d47eaa -> origin/ca_fix_0431d47eaa 2025-09-07T07:36:03.6461478Z * [new branch] camyll/revert-94bc900da97ad7f3c35b3b819bb53b23c74b581a-for-release-2.8 -> origin/camyll/revert-94bc900da97ad7f3c35b3b819bb53b23c74b581a-for-release-2.8 2025-09-07T07:36:03.6463076Z * [new branch] camyllh/test_setup_hooks_push -> origin/camyllh/test_setup_hooks_push 2025-09-07T07:36:03.6464426Z * [new branch] cherry-pick-149654-by-pytorch_bot_bot_ -> origin/cherry-pick-149654-by-pytorch_bot_bot_ 2025-09-07T07:36:03.6465701Z * [new branch] cherry-pick-151939-by-pytorch_bot_bot_ -> origin/cherry-pick-151939-by-pytorch_bot_bot_ 2025-09-07T07:36:03.6467000Z * [new branch] cherry-pick-154174-by-pytorch_bot_bot_ -> origin/cherry-pick-154174-by-pytorch_bot_bot_ 2025-09-07T07:36:03.6468355Z * [new branch] cherry-pick-156260-by-pytorch_bot_bot_ -> origin/cherry-pick-156260-by-pytorch_bot_bot_ 2025-09-07T07:36:03.6469614Z * [new branch] cherry-pick-157453-by-pytorch_bot_bot_ -> origin/cherry-pick-157453-by-pytorch_bot_bot_ 2025-09-07T07:36:03.6470954Z * [new branch] cherry-pick-157513-by-pytorch_bot_bot_ -> origin/cherry-pick-157513-by-pytorch_bot_bot_ 2025-09-07T07:36:03.6472148Z * [new branch] cherry-pick-157695-by-pytorch_bot_bot_ -> origin/cherry-pick-157695-by-pytorch_bot_bot_ 2025-09-07T07:36:03.6473383Z * [new branch] cherry-pick-157732-by-pytorch_bot_bot_ -> origin/cherry-pick-157732-by-pytorch_bot_bot_ 2025-09-07T07:36:03.6474595Z * [new branch] cherry-pick-158537-by-pytorch_bot_bot_ -> origin/cherry-pick-158537-by-pytorch_bot_bot_ 2025-09-07T07:36:03.6475914Z * [new branch] cherry-pick-159969-by-pytorch_bot_bot_ -> origin/cherry-pick-159969-by-pytorch_bot_bot_ 2025-09-07T07:36:03.6477310Z * [new branch] cherry-pick-160586-by-pytorch_bot_bot_ -> origin/cherry-pick-160586-by-pytorch_bot_bot_ 2025-09-07T07:36:03.6479031Z * [new branch] chilli/flex_vllm -> origin/chilli/flex_vllm 2025-09-07T07:36:03.6480363Z * [new branch] cleanup-inductor-benchmark-images -> origin/cleanup-inductor-benchmark-images 2025-09-07T07:36:03.6481525Z * [new branch] codex-testing -> origin/codex-testing 2025-09-07T07:36:03.6483536Z * [new branch] codex/add-helper-function-to-sizevars.py -> origin/codex/add-helper-function-to-sizevars.py 2025-09-07T07:36:03.6484697Z * [new branch] codex/add-helper-function-to-sizevars.py_2025-09-05 -> origin/codex/add-helper-function-to-sizevars.py_2025-09-05 2025-09-07T07:36:03.6485808Z * [new branch] codex/add-metadata-field-for-file-path -> origin/codex/add-metadata-field-for-file-path 2025-09-07T07:36:03.6487537Z * [new branch] codex/add-test-for-inductor-local-cache-behavior -> origin/codex/add-test-for-inductor-local-cache-behavior 2025-09-07T07:36:03.6489400Z * [new branch] codex/create-test-for-tensor-memory-leak-in-cudagraph -> origin/codex/create-test-for-tensor-memory-leak-in-cudagraph 2025-09-07T07:36:03.6490448Z * [new branch] codex/fix-issue-121219-in-pytorch -> origin/codex/fix-issue-121219-in-pytorch 2025-09-07T07:36:03.6491666Z * [new branch] codex/fix-issue-160415-in-pytorch -> origin/codex/fix-issue-160415-in-pytorch 2025-09-07T07:36:03.6492971Z * [new branch] codex/fix-noqengine-quantized-engine-support -> origin/codex/fix-noqengine-quantized-engine-support 2025-09-07T07:36:03.6494139Z * [new branch] codex/fix-pin_memory-error-handling -> origin/codex/fix-pin_memory-error-handling 2025-09-07T07:36:03.6495344Z * [new branch] codex/propose-fix-for-issue-160332 -> origin/codex/propose-fix-for-issue-160332 2025-09-07T07:36:03.6496647Z * [new branch] codex/refactor-lintrunner-config-to-use-uv-run -> origin/codex/refactor-lintrunner-config-to-use-uv-run 2025-09-07T07:36:03.6497906Z * [new branch] codex/remove-allow-untyped-defs-and-fix-type-errors -> origin/codex/remove-allow-untyped-defs-and-fix-type-errors 2025-09-07T07:36:03.6499164Z * [new branch] compile_fsdp2_disable_stream_and_event -> origin/compile_fsdp2_disable_stream_and_event 2025-09-07T07:36:03.6500264Z * [new branch] context_test -> origin/context_test 2025-09-07T07:36:03.6502078Z * [new branch] copilot/fix-157446 -> origin/copilot/fix-157446 2025-09-07T07:36:03.6503536Z * [new branch] copy_graph -> origin/copy_graph 2025-09-07T07:36:03.6505364Z * [new branch] cpio/fix_new_ami_tests -> origin/cpio/fix_new_ami_tests 2025-09-07T07:36:03.6507126Z * [new branch] csl/always_produce_xml -> origin/csl/always_produce_xml 2025-09-07T07:36:03.6508346Z * [new branch] csl/build_test_more_procs -> origin/csl/build_test_more_procs 2025-09-07T07:36:03.6509548Z * [new branch] csl/build_test_more_procs2 -> origin/csl/build_test_more_procs2 2025-09-07T07:36:03.6510732Z * [new branch] csl/disable_flaky_cpp_test -> origin/csl/disable_flaky_cpp_test 2025-09-07T07:36:03.6511877Z * [new branch] csl/disable_periodic_test -> origin/csl/disable_periodic_test 2025-09-07T07:36:03.6513271Z * [new branch] csl/exclude_rocm_viable_strict -> origin/csl/exclude_rocm_viable_strict 2025-09-07T07:36:03.6514793Z * [new branch] csl/katex -> origin/csl/katex 2025-09-07T07:36:03.6516083Z * [new branch] csl/larger_runner -> origin/csl/larger_runner 2025-09-07T07:36:03.6517313Z * [new branch] csl/lintrunner_stuff -> origin/csl/lintrunner_stuff 2025-09-07T07:36:03.6518550Z * [new branch] csl/mps_sharding -> origin/csl/mps_sharding 2025-09-07T07:36:03.6519753Z * [new branch] csl/multistage_docker -> origin/csl/multistage_docker 2025-09-07T07:36:03.6520995Z * [new branch] csl/name_link_check_job -> origin/csl/name_link_check_job 2025-09-07T07:36:03.6522394Z * [new branch] csl/no_keep_goin_rocm -> origin/csl/no_keep_goin_rocm 2025-09-07T07:36:03.6523629Z * [new branch] csl/not_600_timeout -> origin/csl/not_600_timeout 2025-09-07T07:36:03.6524821Z * [new branch] csl/revert_open -> origin/csl/revert_open 2025-09-07T07:36:03.6526089Z * [new branch] csl/skip_build -> origin/csl/skip_build 2025-09-07T07:36:03.6527391Z * [new branch] csl/test_cuda_build_large_runner -> origin/csl/test_cuda_build_large_runner 2025-09-07T07:36:03.6528639Z * [new branch] csl/win_sccache -> origin/csl/win_sccache 2025-09-07T07:36:03.6529892Z * [new branch] cublasltrelax2 -> origin/cublasltrelax2 2025-09-07T07:36:03.6531139Z * [new branch] cublasrelax2 -> origin/cublasrelax2 2025-09-07T07:36:03.6532511Z * [new branch] cudnnsdparefactor -> origin/cudnnsdparefactor 2025-09-07T07:36:03.6533620Z * [new branch] custom_lowering_dict -> origin/custom_lowering_dict 2025-09-07T07:36:03.6535270Z * [new branch] czhuge_muon_dev -> origin/czhuge_muon_dev 2025-09-07T07:36:03.6537141Z * [new branch] d4l3k/delete_hook -> origin/d4l3k/delete_hook 2025-09-07T07:36:03.6538498Z * [new branch] dcp_zoc -> origin/dcp_zoc 2025-09-07T07:36:03.6539801Z * [new branch] debug-guard -> origin/debug-guard 2025-09-07T07:36:03.6541064Z * [new branch] delete-quant-docs -> origin/delete-quant-docs 2025-09-07T07:36:03.6545368Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.2 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.2 2025-09-07T07:36:03.6546783Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.3 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.3 2025-09-07T07:36:03.6548125Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.4 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.55.4 2025-09-07T07:36:03.6549456Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.56.0 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.56.0 2025-09-07T07:36:03.6550814Z * [new branch] dependabot/pip/dot-ci/docker/protobuf-5.29.5 -> origin/dependabot/pip/dot-ci/docker/protobuf-5.29.5 2025-09-07T07:36:03.6553128Z * [new branch] dependabot/pip/dot-github/requirements/protobuf-5.29.5 -> origin/dependabot/pip/dot-github/requirements/protobuf-5.29.5 2025-09-07T07:36:03.6554827Z * [new branch] desertfire/test_cpp_wrapper -> origin/desertfire/test_cpp_wrapper 2025-09-07T07:36:03.6556064Z * [new branch] desertfire/triton-cpu-for-aarch64 -> origin/desertfire/triton-cpu-for-aarch64 2025-09-07T07:36:03.6558527Z * [new branch] dev/joona/MPSNDArrayAdd -> origin/dev/joona/MPSNDArrayAdd 2025-09-07T07:36:03.6559842Z * [new branch] dev/joona/Unranked -> origin/dev/joona/Unranked 2025-09-07T07:36:03.6561443Z * [new branch] dev/joona/cat -> origin/dev/joona/cat 2025-09-07T07:36:03.6562750Z * [new branch] dev/joona/cat_remove_graph -> origin/dev/joona/cat_remove_graph 2025-09-07T07:36:03.6563956Z * [new branch] dev/joona/embeddingbag -> origin/dev/joona/embeddingbag 2025-09-07T07:36:03.6565351Z * [new branch] dev/joona/getTensorsString -> origin/dev/joona/getTensorsString 2025-09-07T07:36:03.6566975Z * [new branch] dev/joona/maxpool2dwithindices_errmsg -> origin/dev/joona/maxpool2dwithindices_errmsg 2025-09-07T07:36:03.6568689Z * [new branch] dev/joona/mps_linear_macos14 -> origin/dev/joona/mps_linear_macos14 2025-09-07T07:36:03.6570452Z * [new branch] dev/joona/sdpa -> origin/dev/joona/sdpa 2025-09-07T07:36:03.6571940Z * [new branch] dev/joona/topk_newapi -> origin/dev/joona/topk_newapi 2025-09-07T07:36:03.6573330Z * [new branch] dev/joona/type_inf -> origin/dev/joona/type_inf 2025-09-07T07:36:03.6574688Z * [new branch] dev/joona/upsize3d -> origin/dev/joona/upsize3d 2025-09-07T07:36:03.6575886Z * [new branch] disable -> origin/disable 2025-09-07T07:36:03.6577086Z * [new branch] e2e-baseline -> origin/e2e-baseline 2025-09-07T07:36:03.6578378Z * [new branch] eigen_for_sparse_addmm_v2 -> origin/eigen_for_sparse_addmm_v2 2025-09-07T07:36:03.6580327Z * [new branch] embg/test_inductor_ci_128B -> origin/embg/test_inductor_ci_128B 2025-09-07T07:36:03.6581607Z * [new branch] embg/test_inductor_ci_base -> origin/embg/test_inductor_ci_base 2025-09-07T07:36:03.6582765Z * [new branch] embg/test_inductor_ci_control -> origin/embg/test_inductor_ci_control 2025-09-07T07:36:03.6583896Z * [new branch] embg/triton_l2_prefetch_128B -> origin/embg/triton_l2_prefetch_128B 2025-09-07T07:36:03.6585294Z * [new branch] embg/triton_l2_prefetch_256B -> origin/embg/triton_l2_prefetch_256B 2025-09-07T07:36:03.6586689Z * [new branch] eqy-patch-1 -> origin/eqy-patch-1 2025-09-07T07:36:03.6587904Z * [new branch] eqy-patch-2 -> origin/eqy-patch-2 2025-09-07T07:36:03.6589083Z * [new branch] eqy-patch-3 -> origin/eqy-patch-3 2025-09-07T07:36:03.6590278Z * [new branch] eqy-patch-4 -> origin/eqy-patch-4 2025-09-07T07:36:03.6591590Z * [new branch] example-convert-torch.nn -> origin/example-convert-torch.nn 2025-09-07T07:36:03.6593541Z * [new branch] exclamaforte/add-contiguous-threshold -> origin/exclamaforte/add-contiguous-threshold 2025-09-07T07:36:03.6594683Z * [new branch] exclamaforte/amd-ma -> origin/exclamaforte/amd-ma 2025-09-07T07:36:03.6596105Z * [new branch] exclamaforte/bump-transformer-version -> origin/exclamaforte/bump-transformer-version 2025-09-07T07:36:03.6597321Z * [new branch] exclamaforte/clear-feedback-savers -> origin/exclamaforte/clear-feedback-savers 2025-09-07T07:36:03.6598505Z * [new branch] exclamaforte/combo-kernels-perf-run -> origin/exclamaforte/combo-kernels-perf-run 2025-09-07T07:36:03.6599996Z * [new branch] exclamaforte/do_bench_refactor -> origin/exclamaforte/do_bench_refactor 2025-09-07T07:36:03.6601719Z * [new branch] exclamaforte/enable-mem-dep-fusion -> origin/exclamaforte/enable-mem-dep-fusion 2025-09-07T07:36:03.6603215Z * [new branch] exclamaforte/fix-exhaustive-autotuning -> origin/exclamaforte/fix-exhaustive-autotuning 2025-09-07T07:36:03.6604693Z * [new branch] exclamaforte/fix-exhuastive-autotuning-reland -> origin/exclamaforte/fix-exhuastive-autotuning-reland 2025-09-07T07:36:03.6605932Z * [new branch] exclamaforte/fix-trace-parsing-fx-svg -> origin/exclamaforte/fix-trace-parsing-fx-svg 2025-09-07T07:36:03.6607189Z * [new branch] exclamaforte/force-pointwise-cat-perf-run -> origin/exclamaforte/force-pointwise-cat-perf-run 2025-09-07T07:36:03.6608468Z * [new branch] exclamaforte/fusion-data -> origin/exclamaforte/fusion-data 2025-09-07T07:36:03.6609722Z * [new branch] exclamaforte/gemm-benchmark-run -> origin/exclamaforte/gemm-benchmark-run 2025-09-07T07:36:03.6611007Z * [new branch] exclamaforte/gemm-export-model -> origin/exclamaforte/gemm-export-model 2025-09-07T07:36:03.6612193Z * [new branch] exclamaforte/gemm-model -> origin/exclamaforte/gemm-model 2025-09-07T07:36:03.6613612Z * [new branch] exclamaforte/gemm-model-all-data-collection -> origin/exclamaforte/gemm-model-all-data-collection 2025-09-07T07:36:03.6614724Z * [new branch] exclamaforte/gemm-to-amd -> origin/exclamaforte/gemm-to-amd 2025-09-07T07:36:03.6615995Z * [new branch] exclamaforte/just-gemm-model -> origin/exclamaforte/just-gemm-model 2025-09-07T07:36:03.6617336Z * [new branch] exclamaforte/just-gemm-model-no-refactor -> origin/exclamaforte/just-gemm-model-no-refactor 2025-09-07T07:36:03.6618551Z * [new branch] exclamaforte/max-autotune-ieee -> origin/exclamaforte/max-autotune-ieee 2025-09-07T07:36:03.6619818Z * [new branch] exclamaforte/memory-counter -> origin/exclamaforte/memory-counter 2025-09-07T07:36:03.6621038Z * [new branch] exclamaforte/profile-diff-algo -> origin/exclamaforte/profile-diff-algo 2025-09-07T07:36:03.6622243Z * [new branch] exclamaforte/profiler-combo -> origin/exclamaforte/profiler-combo 2025-09-07T07:36:03.6623693Z * [new branch] exclamaforte/test_cpp_wrapper_mode -> origin/exclamaforte/test_cpp_wrapper_mode 2025-09-07T07:36:03.6624867Z * [new branch] exclamaforte/update-autotune-configs -> origin/exclamaforte/update-autotune-configs 2025-09-07T07:36:03.6626076Z * [new branch] exclamaforte/update-autotune-configs-2 -> origin/exclamaforte/update-autotune-configs-2 2025-09-07T07:36:03.6627857Z * [new branch] exclamforte/gemm-model-final -> origin/exclamforte/gemm-model-final 2025-09-07T07:36:03.6629217Z * [new branch] exec -> origin/exec 2025-09-07T07:36:03.6630456Z * [new branch] executorch-module-shim -> origin/executorch-module-shim 2025-09-07T07:36:03.6631797Z * [new branch] experimental-mosaic -> origin/experimental-mosaic 2025-09-07T07:36:03.6633040Z * [new branch] export-D58091437 -> origin/export-D58091437 2025-09-07T07:36:03.6634357Z * [new branch] export-D61047529 -> origin/export-D61047529 2025-09-07T07:36:03.6635561Z * [new branch] export-D70112642 -> origin/export-D70112642 2025-09-07T07:36:03.6636846Z * [new branch] export-D71412006 -> origin/export-D71412006 2025-09-07T07:36:03.6638246Z * [new branch] export-D73042989 -> origin/export-D73042989 2025-09-07T07:36:03.6639513Z * [new branch] export-D75183591 -> origin/export-D75183591 2025-09-07T07:36:03.6640768Z * [new branch] export-D75617432 -> origin/export-D75617432 2025-09-07T07:36:03.6641972Z * [new branch] export-D75659965 -> origin/export-D75659965 2025-09-07T07:36:03.6643182Z * [new branch] export-D76080931 -> origin/export-D76080931 2025-09-07T07:36:03.6644405Z * [new branch] export-D76797250 -> origin/export-D76797250 2025-09-07T07:36:03.6645580Z * [new branch] export-D76885271 -> origin/export-D76885271 2025-09-07T07:36:03.6646770Z * [new branch] export-D76885620 -> origin/export-D76885620 2025-09-07T07:36:03.6648040Z * [new branch] export-D76936623 -> origin/export-D76936623 2025-09-07T07:36:03.6649319Z * [new branch] export-D76958268 -> origin/export-D76958268 2025-09-07T07:36:03.6650571Z * [new branch] export-D78375400 -> origin/export-D78375400 2025-09-07T07:36:03.6651847Z * [new branch] export-D78431305 -> origin/export-D78431305 2025-09-07T07:36:03.6653093Z * [new branch] export-D78580107 -> origin/export-D78580107 2025-09-07T07:36:03.6654318Z * [new branch] export-D78822171 -> origin/export-D78822171 2025-09-07T07:36:03.6655571Z * [new branch] export-D78822351 -> origin/export-D78822351 2025-09-07T07:36:03.6656702Z * [new branch] export-D78822507 -> origin/export-D78822507 2025-09-07T07:36:03.6657991Z * [new branch] export-D78826994 -> origin/export-D78826994 2025-09-07T07:36:03.6659245Z * [new branch] export-D78894324 -> origin/export-D78894324 2025-09-07T07:36:03.6660570Z * [new branch] export-D78929245 -> origin/export-D78929245 2025-09-07T07:36:03.6661778Z * [new branch] export-D78934925 -> origin/export-D78934925 2025-09-07T07:36:03.6663069Z * [new branch] export-D78953203 -> origin/export-D78953203 2025-09-07T07:36:03.6664276Z * [new branch] export-D78953229 -> origin/export-D78953229 2025-09-07T07:36:03.6665410Z * [new branch] export-D78957093 -> origin/export-D78957093 2025-09-07T07:36:03.6666614Z * [new branch] export-D78957389 -> origin/export-D78957389 2025-09-07T07:36:03.6667853Z * [new branch] export-D78996107 -> origin/export-D78996107 2025-09-07T07:36:03.6669073Z * [new branch] export-D79026433 -> origin/export-D79026433 2025-09-07T07:36:03.6670273Z * [new branch] export-D79230339 -> origin/export-D79230339 2025-09-07T07:36:03.6671474Z * [new branch] export-D79319835 -> origin/export-D79319835 2025-09-07T07:36:03.6672681Z * [new branch] export-D79328456 -> origin/export-D79328456 2025-09-07T07:36:03.6673906Z * [new branch] export-D79534608 -> origin/export-D79534608 2025-09-07T07:36:03.6675309Z * [new branch] export-D79785974 -> origin/export-D79785974 2025-09-07T07:36:03.6676560Z * [new branch] export-D80025417 -> origin/export-D80025417 2025-09-07T07:36:03.6677763Z * [new branch] export-D80120333 -> origin/export-D80120333 2025-09-07T07:36:03.6679117Z * [new branch] export-D80214882 -> origin/export-D80214882 2025-09-07T07:36:03.6680307Z * [new branch] export-D80319069 -> origin/export-D80319069 2025-09-07T07:36:03.6681607Z * [new branch] export-D80321215 -> origin/export-D80321215 2025-09-07T07:36:03.6682796Z * [new branch] export-D80503451 -> origin/export-D80503451 2025-09-07T07:36:03.6683972Z * [new branch] export-D80771648 -> origin/export-D80771648 2025-09-07T07:36:03.6685159Z * [new branch] export-D80823877 -> origin/export-D80823877 2025-09-07T07:36:03.6686422Z * [new branch] export-D80948073 -> origin/export-D80948073 2025-09-07T07:36:03.6687702Z * [new branch] export-D80958642 -> origin/export-D80958642 2025-09-07T07:36:03.6688995Z * [new branch] export-D80970483 -> origin/export-D80970483 2025-09-07T07:36:03.6690190Z * [new branch] export-D81054193 -> origin/export-D81054193 2025-09-07T07:36:03.6691511Z * [new branch] export-D81060182 -> origin/export-D81060182 2025-09-07T07:36:03.6693215Z * [new branch] export-D81078973 -> origin/export-D81078973 2025-09-07T07:36:03.6694421Z * [new branch] export-D81204584 -> origin/export-D81204584 2025-09-07T07:36:03.6695676Z * [new branch] export-D81284190 -> origin/export-D81284190 2025-09-07T07:36:03.6696958Z * [new branch] export-D81299840 -> origin/export-D81299840 2025-09-07T07:36:03.6698237Z * [new branch] export-D81429090 -> origin/export-D81429090 2025-09-07T07:36:03.6699483Z * [new branch] export-D81698719 -> origin/export-D81698719 2025-09-07T07:36:03.6700717Z * [new branch] export-D81747409 -> origin/export-D81747409 2025-09-07T07:36:03.6702140Z * [new branch] exported-model-train-idempotent -> origin/exported-model-train-idempotent 2025-09-07T07:36:03.6704070Z * [new branch] ezyang/wip-aot-descriptors -> origin/ezyang/wip-aot-descriptors 2025-09-07T07:36:03.6705662Z * [new branch] fa_u8_brgemm -> origin/fa_u8_brgemm 2025-09-07T07:36:03.6706967Z * [new branch] fastmath_baseline -> origin/fastmath_baseline 2025-09-07T07:36:03.6708779Z * [new branch] fbcode/warm -> origin/fbcode/warm 2025-09-07T07:36:03.6710102Z * [new branch] fca -> origin/fca 2025-09-07T07:36:03.6711272Z * [new branch] fca2_ca5984c -> origin/fca2_ca5984c 2025-09-07T07:36:03.6712515Z * [new branch] fca5 -> origin/fca5 2025-09-07T07:36:03.6714361Z * [new branch] feature/function-numa-binding -> origin/feature/function-numa-binding 2025-09-07T07:36:03.6715666Z * [new branch] feature/function-numa-binding-take2 -> origin/feature/function-numa-binding-take2 2025-09-07T07:36:03.6716584Z * [new branch] feature/numa-nproc-fix -> origin/feature/numa-nproc-fix 2025-09-07T07:36:03.6717816Z * [new branch] feature/numa-signpost-serialize -> origin/feature/numa-signpost-serialize 2025-09-07T07:36:03.6718945Z * [new branch] feature/parallel-numa-binding -> origin/feature/parallel-numa-binding 2025-09-07T07:36:03.6720748Z * [new branch] fengyuan/external-proj -> origin/fengyuan/external-proj 2025-09-07T07:36:03.6721986Z * [new branch] fengyuan/out-of-tree-xpu-ops-improve-test -> origin/fengyuan/out-of-tree-xpu-ops-improve-test 2025-09-07T07:36:03.6723081Z * [new branch] fengyuan/out-of-tree-xpu-ops-remove-dtype -> origin/fengyuan/out-of-tree-xpu-ops-remove-dtype 2025-09-07T07:36:03.6724175Z * [new branch] fengyuan/test-xpu -> origin/fengyuan/test-xpu 2025-09-07T07:36:03.6725782Z * [new branch] ffast_math_baseline -> origin/ffast_math_baseline 2025-09-07T07:36:03.6727142Z * [new branch] ffast_math_target -> origin/ffast_math_target 2025-09-07T07:36:03.6728936Z * [new branch] findhao/base_commit -> origin/findhao/base_commit 2025-09-07T07:36:03.6730105Z * [new branch] findhao/base_commit1 -> origin/findhao/base_commit1 2025-09-07T07:36:03.6731375Z * [new branch] findhao/multistream2 -> origin/findhao/multistream2 2025-09-07T07:36:03.6732489Z * [new branch] findhao/multistream5 -> origin/findhao/multistream5 2025-09-07T07:36:03.6733678Z * [new branch] findhao/multistream6 -> origin/findhao/multistream6 2025-09-07T07:36:03.6734855Z * [new branch] findhao/operatorbench3 -> origin/findhao/operatorbench3 2025-09-07T07:36:03.6736014Z * [new branch] findhao/operatorbench5 -> origin/findhao/operatorbench5 2025-09-07T07:36:03.6737155Z * [new branch] findhao/tritonparse -> origin/findhao/tritonparse 2025-09-07T07:36:03.6738463Z * [new branch] fix -> origin/fix 2025-09-07T07:36:03.6739790Z * [new branch] fix-ck-gemm-template-format -> origin/fix-ck-gemm-template-format 2025-09-07T07:36:03.6740973Z * [new branch] fix-config-ignore -> origin/fix-config-ignore 2025-09-07T07:36:03.6757708Z * [new branch] fix-dict-guard -> origin/fix-dict-guard 2025-09-07T07:36:03.6758314Z * [new branch] fix-inductor-periodic-0528 -> origin/fix-inductor-periodic-0528 2025-09-07T07:36:03.6758692Z * [new branch] fix-mps-benchmark -> origin/fix-mps-benchmark 2025-09-07T07:36:03.6759072Z * [new branch] fix-rlease-feature-template -> origin/fix-rlease-feature-template 2025-09-07T07:36:03.6759516Z * [new branch] fix-run-condition-upload-results -> origin/fix-run-condition-upload-results 2025-09-07T07:36:03.6759907Z * [new branch] fix-torchbench -> origin/fix-torchbench 2025-09-07T07:36:03.6760202Z * [new branch] fix_153389 -> origin/fix_153389 2025-09-07T07:36:03.6760510Z * [new branch] fix_fsdp_rs_bucket2 -> origin/fix_fsdp_rs_bucket2 2025-09-07T07:36:03.6760853Z * [new branch] fix_inductor_peridic_tests -> origin/fix_inductor_peridic_tests 2025-09-07T07:36:03.6761181Z * [new branch] fix_ubn_159469 -> origin/fix_ubn_159469 2025-09-07T07:36:03.6761475Z * [new branch] fixes-triage -> origin/fixes-triage 2025-09-07T07:36:03.6761771Z * [new branch] fixflashinfer -> origin/fixflashinfer 2025-09-07T07:36:03.6762076Z * [new branch] flash_decoding_cpu -> origin/flash_decoding_cpu 2025-09-07T07:36:03.6762494Z * [new branch] flex-flash -> origin/flex-flash 2025-09-07T07:36:03.6762800Z * [new branch] flex-lowering -> origin/flex-lowering 2025-09-07T07:36:03.6763092Z * [new branch] flex-warning -> origin/flex-warning 2025-09-07T07:36:03.6763432Z * [new branch] flex_attention_functorch_grad -> origin/flex_attention_functorch_grad 2025-09-07T07:36:03.6763768Z * [new branch] flex_flash -> origin/flex_flash 2025-09-07T07:36:03.6764402Z * [new branch] flexdecode-gqa-groups -> origin/flexdecode-gqa-groups 2025-09-07T07:36:03.6766295Z * [new branch] fmassa/fix_memeff_sharding_rule -> origin/fmassa/fix_memeff_sharding_rule 2025-09-07T07:36:03.6767496Z * [new branch] fsdp2_trace_rules -> origin/fsdp2_trace_rules 2025-09-07T07:36:03.6768745Z * [new branch] fsdpv2_3d -> origin/fsdpv2_3d 2025-09-07T07:36:03.6770174Z * [new branch] fsdpv2_3d_m1 -> origin/fsdpv2_3d_m1 2025-09-07T07:36:03.6771452Z * [new branch] fx_cpp -> origin/fx_cpp 2025-09-07T07:36:03.6773308Z * [new branch] fy/fix-win -> origin/fy/fix-win 2025-09-07T07:36:03.6776336Z * [new branch] gh/AlnisM/1/base -> origin/gh/AlnisM/1/base 2025-09-07T07:36:03.6777504Z * [new branch] gh/AlnisM/1/head -> origin/gh/AlnisM/1/head 2025-09-07T07:36:03.6779751Z * [new branch] gh/CaoE/2/base -> origin/gh/CaoE/2/base 2025-09-07T07:36:03.6780949Z * [new branch] gh/CaoE/2/head -> origin/gh/CaoE/2/head 2025-09-07T07:36:03.6782170Z * [new branch] gh/CaoE/2/orig -> origin/gh/CaoE/2/orig 2025-09-07T07:36:03.6784442Z * [new branch] gh/ColinPeppler/79/base -> origin/gh/ColinPeppler/79/base 2025-09-07T07:36:03.6785709Z * [new branch] gh/ColinPeppler/79/head -> origin/gh/ColinPeppler/79/head 2025-09-07T07:36:03.6786911Z * [new branch] gh/ColinPeppler/79/orig -> origin/gh/ColinPeppler/79/orig 2025-09-07T07:36:03.6788808Z * [new branch] gh/ColinPeppler/80/base -> origin/gh/ColinPeppler/80/base 2025-09-07T07:36:03.6790134Z * [new branch] gh/ColinPeppler/80/head -> origin/gh/ColinPeppler/80/head 2025-09-07T07:36:03.6791314Z * [new branch] gh/ColinPeppler/80/orig -> origin/gh/ColinPeppler/80/orig 2025-09-07T07:36:03.6793588Z * [new branch] gh/EikanWang/67/base -> origin/gh/EikanWang/67/base 2025-09-07T07:36:03.6794754Z * [new branch] gh/EikanWang/67/head -> origin/gh/EikanWang/67/head 2025-09-07T07:36:03.6796744Z * [new branch] gh/EikanWang/80/base -> origin/gh/EikanWang/80/base 2025-09-07T07:36:03.6797964Z * [new branch] gh/EikanWang/80/head -> origin/gh/EikanWang/80/head 2025-09-07T07:36:03.6799145Z * [new branch] gh/EikanWang/80/orig -> origin/gh/EikanWang/80/orig 2025-09-07T07:36:03.6800860Z * [new branch] gh/EikanWang/81/base -> origin/gh/EikanWang/81/base 2025-09-07T07:36:03.6802068Z * [new branch] gh/EikanWang/81/head -> origin/gh/EikanWang/81/head 2025-09-07T07:36:03.6803435Z * [new branch] gh/EikanWang/81/orig -> origin/gh/EikanWang/81/orig 2025-09-07T07:36:03.6805447Z * [new branch] gh/EikanWang/82/base -> origin/gh/EikanWang/82/base 2025-09-07T07:36:03.6806683Z * [new branch] gh/EikanWang/82/head -> origin/gh/EikanWang/82/head 2025-09-07T07:36:03.6807927Z * [new branch] gh/EikanWang/82/orig -> origin/gh/EikanWang/82/orig 2025-09-07T07:36:03.6810330Z * [new branch] gh/Gasoonjia/1/base -> origin/gh/Gasoonjia/1/base 2025-09-07T07:36:03.6811551Z * [new branch] gh/Gasoonjia/1/head -> origin/gh/Gasoonjia/1/head 2025-09-07T07:36:03.6813995Z * [new branch] gh/H-Huang/131/base -> origin/gh/H-Huang/131/base 2025-09-07T07:36:03.6815023Z * [new branch] gh/H-Huang/131/head -> origin/gh/H-Huang/131/head 2025-09-07T07:36:03.6816224Z * [new branch] gh/H-Huang/131/orig -> origin/gh/H-Huang/131/orig 2025-09-07T07:36:03.6818110Z * [new branch] gh/H-Huang/132/base -> origin/gh/H-Huang/132/base 2025-09-07T07:36:03.6819423Z * [new branch] gh/H-Huang/132/head -> origin/gh/H-Huang/132/head 2025-09-07T07:36:03.6820612Z * [new branch] gh/H-Huang/132/orig -> origin/gh/H-Huang/132/orig 2025-09-07T07:36:03.6822439Z * [new branch] gh/H-Huang/180/base -> origin/gh/H-Huang/180/base 2025-09-07T07:36:03.6823608Z * [new branch] gh/H-Huang/180/head -> origin/gh/H-Huang/180/head 2025-09-07T07:36:03.6824860Z * [new branch] gh/H-Huang/180/orig -> origin/gh/H-Huang/180/orig 2025-09-07T07:36:03.6826542Z * [new branch] gh/H-Huang/182/base -> origin/gh/H-Huang/182/base 2025-09-07T07:36:03.6827694Z * [new branch] gh/H-Huang/182/head -> origin/gh/H-Huang/182/head 2025-09-07T07:36:03.6828929Z * [new branch] gh/H-Huang/182/orig -> origin/gh/H-Huang/182/orig 2025-09-07T07:36:03.6830752Z * [new branch] gh/H-Huang/187/base -> origin/gh/H-Huang/187/base 2025-09-07T07:36:03.6831907Z * [new branch] gh/H-Huang/187/head -> origin/gh/H-Huang/187/head 2025-09-07T07:36:03.6833190Z * [new branch] gh/H-Huang/187/orig -> origin/gh/H-Huang/187/orig 2025-09-07T07:36:03.6835016Z * [new branch] gh/H-Huang/202/base -> origin/gh/H-Huang/202/base 2025-09-07T07:36:03.6836199Z * [new branch] gh/H-Huang/202/head -> origin/gh/H-Huang/202/head 2025-09-07T07:36:03.6837373Z * [new branch] gh/H-Huang/202/orig -> origin/gh/H-Huang/202/orig 2025-09-07T07:36:03.6839079Z * [new branch] gh/H-Huang/203/base -> origin/gh/H-Huang/203/base 2025-09-07T07:36:03.6840299Z * [new branch] gh/H-Huang/203/head -> origin/gh/H-Huang/203/head 2025-09-07T07:36:03.6841559Z * [new branch] gh/H-Huang/203/orig -> origin/gh/H-Huang/203/orig 2025-09-07T07:36:03.6843351Z * [new branch] gh/H-Huang/204/base -> origin/gh/H-Huang/204/base 2025-09-07T07:36:03.6844586Z * [new branch] gh/H-Huang/204/head -> origin/gh/H-Huang/204/head 2025-09-07T07:36:03.6845769Z * [new branch] gh/H-Huang/204/orig -> origin/gh/H-Huang/204/orig 2025-09-07T07:36:03.6847476Z * [new branch] gh/H-Huang/205/base -> origin/gh/H-Huang/205/base 2025-09-07T07:36:03.6848767Z * [new branch] gh/H-Huang/205/head -> origin/gh/H-Huang/205/head 2025-09-07T07:36:03.6849975Z * [new branch] gh/H-Huang/205/orig -> origin/gh/H-Huang/205/orig 2025-09-07T07:36:03.6851668Z * [new branch] gh/H-Huang/206/base -> origin/gh/H-Huang/206/base 2025-09-07T07:36:03.6852894Z * [new branch] gh/H-Huang/206/head -> origin/gh/H-Huang/206/head 2025-09-07T07:36:03.6854080Z * [new branch] gh/H-Huang/206/orig -> origin/gh/H-Huang/206/orig 2025-09-07T07:36:03.6855816Z * [new branch] gh/H-Huang/207/base -> origin/gh/H-Huang/207/base 2025-09-07T07:36:03.6857018Z * [new branch] gh/H-Huang/207/head -> origin/gh/H-Huang/207/head 2025-09-07T07:36:03.6858381Z * [new branch] gh/H-Huang/207/orig -> origin/gh/H-Huang/207/orig 2025-09-07T07:36:03.6860073Z * [new branch] gh/H-Huang/208/base -> origin/gh/H-Huang/208/base 2025-09-07T07:36:03.6861285Z * [new branch] gh/H-Huang/208/head -> origin/gh/H-Huang/208/head 2025-09-07T07:36:03.6862661Z * [new branch] gh/H-Huang/208/orig -> origin/gh/H-Huang/208/orig 2025-09-07T07:36:03.6864288Z * [new branch] gh/H-Huang/209/base -> origin/gh/H-Huang/209/base 2025-09-07T07:36:03.6865466Z * [new branch] gh/H-Huang/209/head -> origin/gh/H-Huang/209/head 2025-09-07T07:36:03.6866644Z * [new branch] gh/H-Huang/209/orig -> origin/gh/H-Huang/209/orig 2025-09-07T07:36:03.6868481Z * [new branch] gh/H-Huang/210/base -> origin/gh/H-Huang/210/base 2025-09-07T07:36:03.6869652Z * [new branch] gh/H-Huang/210/head -> origin/gh/H-Huang/210/head 2025-09-07T07:36:03.6870841Z * [new branch] gh/H-Huang/210/orig -> origin/gh/H-Huang/210/orig 2025-09-07T07:36:03.6872571Z * [new branch] gh/H-Huang/211/base -> origin/gh/H-Huang/211/base 2025-09-07T07:36:03.6873770Z * [new branch] gh/H-Huang/211/head -> origin/gh/H-Huang/211/head 2025-09-07T07:36:03.6874962Z * [new branch] gh/H-Huang/211/orig -> origin/gh/H-Huang/211/orig 2025-09-07T07:36:03.6876656Z * [new branch] gh/H-Huang/212/base -> origin/gh/H-Huang/212/base 2025-09-07T07:36:03.6877831Z * [new branch] gh/H-Huang/212/head -> origin/gh/H-Huang/212/head 2025-09-07T07:36:03.6879034Z * [new branch] gh/H-Huang/212/orig -> origin/gh/H-Huang/212/orig 2025-09-07T07:36:03.6881726Z * [new branch] gh/H-Huang/213/base -> origin/gh/H-Huang/213/base 2025-09-07T07:36:03.6882981Z * [new branch] gh/H-Huang/213/head -> origin/gh/H-Huang/213/head 2025-09-07T07:36:03.6884138Z * [new branch] gh/H-Huang/213/orig -> origin/gh/H-Huang/213/orig 2025-09-07T07:36:03.6885912Z * [new branch] gh/H-Huang/214/base -> origin/gh/H-Huang/214/base 2025-09-07T07:36:03.6887130Z * [new branch] gh/H-Huang/214/head -> origin/gh/H-Huang/214/head 2025-09-07T07:36:03.6888316Z * [new branch] gh/H-Huang/214/orig -> origin/gh/H-Huang/214/orig 2025-09-07T07:36:03.6890460Z * [new branch] gh/IvanKobzarev/112/base -> origin/gh/IvanKobzarev/112/base 2025-09-07T07:36:03.6891711Z * [new branch] gh/IvanKobzarev/112/head -> origin/gh/IvanKobzarev/112/head 2025-09-07T07:36:03.6892922Z * [new branch] gh/IvanKobzarev/112/orig -> origin/gh/IvanKobzarev/112/orig 2025-09-07T07:36:03.6894740Z * [new branch] gh/IvanKobzarev/115/base -> origin/gh/IvanKobzarev/115/base 2025-09-07T07:36:03.6895983Z * [new branch] gh/IvanKobzarev/115/head -> origin/gh/IvanKobzarev/115/head 2025-09-07T07:36:03.6897265Z * [new branch] gh/IvanKobzarev/115/orig -> origin/gh/IvanKobzarev/115/orig 2025-09-07T07:36:03.6899347Z * [new branch] gh/IvanKobzarev/116/base -> origin/gh/IvanKobzarev/116/base 2025-09-07T07:36:03.6900665Z * [new branch] gh/IvanKobzarev/116/head -> origin/gh/IvanKobzarev/116/head 2025-09-07T07:36:03.6901843Z * [new branch] gh/IvanKobzarev/116/orig -> origin/gh/IvanKobzarev/116/orig 2025-09-07T07:36:03.6903908Z * [new branch] gh/IvanKobzarev/118/base -> origin/gh/IvanKobzarev/118/base 2025-09-07T07:36:03.6905261Z * [new branch] gh/IvanKobzarev/118/head -> origin/gh/IvanKobzarev/118/head 2025-09-07T07:36:03.6906489Z * [new branch] gh/IvanKobzarev/118/orig -> origin/gh/IvanKobzarev/118/orig 2025-09-07T07:36:03.6908428Z * [new branch] gh/IvanKobzarev/126/base -> origin/gh/IvanKobzarev/126/base 2025-09-07T07:36:03.6909671Z * [new branch] gh/IvanKobzarev/126/head -> origin/gh/IvanKobzarev/126/head 2025-09-07T07:36:03.6910878Z * [new branch] gh/IvanKobzarev/126/orig -> origin/gh/IvanKobzarev/126/orig 2025-09-07T07:36:03.6912723Z * [new branch] gh/IvanKobzarev/127/base -> origin/gh/IvanKobzarev/127/base 2025-09-07T07:36:03.6914060Z * [new branch] gh/IvanKobzarev/127/head -> origin/gh/IvanKobzarev/127/head 2025-09-07T07:36:03.6915118Z * [new branch] gh/IvanKobzarev/127/orig -> origin/gh/IvanKobzarev/127/orig 2025-09-07T07:36:03.6916946Z * [new branch] gh/IvanKobzarev/128/base -> origin/gh/IvanKobzarev/128/base 2025-09-07T07:36:03.6918180Z * [new branch] gh/IvanKobzarev/128/head -> origin/gh/IvanKobzarev/128/head 2025-09-07T07:36:03.6919338Z * [new branch] gh/IvanKobzarev/128/orig -> origin/gh/IvanKobzarev/128/orig 2025-09-07T07:36:03.6921179Z * [new branch] gh/IvanKobzarev/132/base -> origin/gh/IvanKobzarev/132/base 2025-09-07T07:36:03.6922432Z * [new branch] gh/IvanKobzarev/132/head -> origin/gh/IvanKobzarev/132/head 2025-09-07T07:36:03.6923661Z * [new branch] gh/IvanKobzarev/132/orig -> origin/gh/IvanKobzarev/132/orig 2025-09-07T07:36:03.6925866Z * [new branch] gh/IvanKobzarev/133/base -> origin/gh/IvanKobzarev/133/base 2025-09-07T07:36:03.6927296Z * [new branch] gh/IvanKobzarev/133/head -> origin/gh/IvanKobzarev/133/head 2025-09-07T07:36:03.6928570Z * [new branch] gh/IvanKobzarev/133/orig -> origin/gh/IvanKobzarev/133/orig 2025-09-07T07:36:03.6930269Z * [new branch] gh/IvanKobzarev/134/base -> origin/gh/IvanKobzarev/134/base 2025-09-07T07:36:03.6931462Z * [new branch] gh/IvanKobzarev/134/head -> origin/gh/IvanKobzarev/134/head 2025-09-07T07:36:03.6932617Z * [new branch] gh/IvanKobzarev/134/orig -> origin/gh/IvanKobzarev/134/orig 2025-09-07T07:36:03.6934615Z * [new branch] gh/IvanKobzarev/135/base -> origin/gh/IvanKobzarev/135/base 2025-09-07T07:36:03.6935811Z * [new branch] gh/IvanKobzarev/135/head -> origin/gh/IvanKobzarev/135/head 2025-09-07T07:36:03.6937027Z * [new branch] gh/IvanKobzarev/135/orig -> origin/gh/IvanKobzarev/135/orig 2025-09-07T07:36:03.6938929Z * [new branch] gh/IvanKobzarev/136/base -> origin/gh/IvanKobzarev/136/base 2025-09-07T07:36:03.6940145Z * [new branch] gh/IvanKobzarev/136/head -> origin/gh/IvanKobzarev/136/head 2025-09-07T07:36:03.6941392Z * [new branch] gh/IvanKobzarev/136/orig -> origin/gh/IvanKobzarev/136/orig 2025-09-07T07:36:03.6943183Z * [new branch] gh/IvanKobzarev/137/base -> origin/gh/IvanKobzarev/137/base 2025-09-07T07:36:03.6944376Z * [new branch] gh/IvanKobzarev/137/head -> origin/gh/IvanKobzarev/137/head 2025-09-07T07:36:03.6945574Z * [new branch] gh/IvanKobzarev/137/orig -> origin/gh/IvanKobzarev/137/orig 2025-09-07T07:36:03.6947311Z * [new branch] gh/IvanKobzarev/138/base -> origin/gh/IvanKobzarev/138/base 2025-09-07T07:36:03.6948526Z * [new branch] gh/IvanKobzarev/138/head -> origin/gh/IvanKobzarev/138/head 2025-09-07T07:36:03.6949847Z * [new branch] gh/IvanKobzarev/138/orig -> origin/gh/IvanKobzarev/138/orig 2025-09-07T07:36:03.6951611Z * [new branch] gh/IvanKobzarev/139/base -> origin/gh/IvanKobzarev/139/base 2025-09-07T07:36:03.6952888Z * [new branch] gh/IvanKobzarev/139/head -> origin/gh/IvanKobzarev/139/head 2025-09-07T07:36:03.6954089Z * [new branch] gh/IvanKobzarev/139/orig -> origin/gh/IvanKobzarev/139/orig 2025-09-07T07:36:03.6955988Z * [new branch] gh/IvanKobzarev/140/base -> origin/gh/IvanKobzarev/140/base 2025-09-07T07:36:03.6957137Z * [new branch] gh/IvanKobzarev/140/head -> origin/gh/IvanKobzarev/140/head 2025-09-07T07:36:03.6958361Z * [new branch] gh/IvanKobzarev/140/orig -> origin/gh/IvanKobzarev/140/orig 2025-09-07T07:36:03.6960671Z * [new branch] gh/IvanKobzarev/141/base -> origin/gh/IvanKobzarev/141/base 2025-09-07T07:36:03.6962105Z * [new branch] gh/IvanKobzarev/141/head -> origin/gh/IvanKobzarev/141/head 2025-09-07T07:36:03.6964418Z * [new branch] gh/IvanKobzarev/141/orig -> origin/gh/IvanKobzarev/141/orig 2025-09-07T07:36:03.6966197Z * [new branch] gh/IvanKobzarev/142/base -> origin/gh/IvanKobzarev/142/base 2025-09-07T07:36:03.6966540Z * [new branch] gh/IvanKobzarev/142/head -> origin/gh/IvanKobzarev/142/head 2025-09-07T07:36:03.6967704Z * [new branch] gh/IvanKobzarev/142/orig -> origin/gh/IvanKobzarev/142/orig 2025-09-07T07:36:03.6969551Z * [new branch] gh/IvanKobzarev/143/base -> origin/gh/IvanKobzarev/143/base 2025-09-07T07:36:03.6970844Z * [new branch] gh/IvanKobzarev/143/head -> origin/gh/IvanKobzarev/143/head 2025-09-07T07:36:03.6972053Z * [new branch] gh/IvanKobzarev/143/orig -> origin/gh/IvanKobzarev/143/orig 2025-09-07T07:36:03.6973952Z * [new branch] gh/IvanKobzarev/144/base -> origin/gh/IvanKobzarev/144/base 2025-09-07T07:36:03.6975147Z * [new branch] gh/IvanKobzarev/144/head -> origin/gh/IvanKobzarev/144/head 2025-09-07T07:36:03.6976320Z * [new branch] gh/IvanKobzarev/144/orig -> origin/gh/IvanKobzarev/144/orig 2025-09-07T07:36:03.6978327Z * [new branch] gh/IvanKobzarev/145/base -> origin/gh/IvanKobzarev/145/base 2025-09-07T07:36:03.6979781Z * [new branch] gh/IvanKobzarev/145/head -> origin/gh/IvanKobzarev/145/head 2025-09-07T07:36:03.6981046Z * [new branch] gh/IvanKobzarev/145/orig -> origin/gh/IvanKobzarev/145/orig 2025-09-07T07:36:03.6982834Z * [new branch] gh/IvanKobzarev/146/base -> origin/gh/IvanKobzarev/146/base 2025-09-07T07:36:03.6984049Z * [new branch] gh/IvanKobzarev/146/head -> origin/gh/IvanKobzarev/146/head 2025-09-07T07:36:03.6985255Z * [new branch] gh/IvanKobzarev/146/orig -> origin/gh/IvanKobzarev/146/orig 2025-09-07T07:36:03.6987401Z * [new branch] gh/NikhilAPatel/1/base -> origin/gh/NikhilAPatel/1/base 2025-09-07T07:36:03.6988694Z * [new branch] gh/NikhilAPatel/1/head -> origin/gh/NikhilAPatel/1/head 2025-09-07T07:36:03.6990343Z * [new branch] gh/NikhilAPatel/2/base -> origin/gh/NikhilAPatel/2/base 2025-09-07T07:36:03.6991511Z * [new branch] gh/NikhilAPatel/2/head -> origin/gh/NikhilAPatel/2/head 2025-09-07T07:36:03.6993418Z * [new branch] gh/NikhilAPatel/4/base -> origin/gh/NikhilAPatel/4/base 2025-09-07T07:36:03.6994759Z * [new branch] gh/NikhilAPatel/4/head -> origin/gh/NikhilAPatel/4/head 2025-09-07T07:36:03.6996793Z * [new branch] gh/PaliC/1/base -> origin/gh/PaliC/1/base 2025-09-07T07:36:03.6998028Z * [new branch] gh/PaliC/1/head -> origin/gh/PaliC/1/head 2025-09-07T07:36:03.6999236Z * [new branch] gh/PaliC/1/orig -> origin/gh/PaliC/1/orig 2025-09-07T07:36:03.7001036Z * [new branch] gh/PaliC/17/base -> origin/gh/PaliC/17/base 2025-09-07T07:36:03.7002211Z * [new branch] gh/PaliC/17/head -> origin/gh/PaliC/17/head 2025-09-07T07:36:03.7003702Z * [new branch] gh/PaliC/17/orig -> origin/gh/PaliC/17/orig 2025-09-07T07:36:03.7005388Z * [new branch] gh/PaliC/18/base -> origin/gh/PaliC/18/base 2025-09-07T07:36:03.7006585Z * [new branch] gh/PaliC/18/head -> origin/gh/PaliC/18/head 2025-09-07T07:36:03.7007769Z * [new branch] gh/PaliC/18/orig -> origin/gh/PaliC/18/orig 2025-09-07T07:36:03.7009455Z * [new branch] gh/PaliC/2/base -> origin/gh/PaliC/2/base 2025-09-07T07:36:03.7010695Z * [new branch] gh/PaliC/2/head -> origin/gh/PaliC/2/head 2025-09-07T07:36:03.7011926Z * [new branch] gh/PaliC/2/orig -> origin/gh/PaliC/2/orig 2025-09-07T07:36:03.7013718Z * [new branch] gh/PaliC/20/base -> origin/gh/PaliC/20/base 2025-09-07T07:36:03.7015051Z * [new branch] gh/PaliC/20/head -> origin/gh/PaliC/20/head 2025-09-07T07:36:03.7016240Z * [new branch] gh/PaliC/20/orig -> origin/gh/PaliC/20/orig 2025-09-07T07:36:03.7018024Z * [new branch] gh/PaliC/21/base -> origin/gh/PaliC/21/base 2025-09-07T07:36:03.7019238Z * [new branch] gh/PaliC/21/head -> origin/gh/PaliC/21/head 2025-09-07T07:36:03.7020454Z * [new branch] gh/PaliC/21/orig -> origin/gh/PaliC/21/orig 2025-09-07T07:36:03.7022142Z * [new branch] gh/PaliC/22/base -> origin/gh/PaliC/22/base 2025-09-07T07:36:03.7023355Z * [new branch] gh/PaliC/22/head -> origin/gh/PaliC/22/head 2025-09-07T07:36:03.7024602Z * [new branch] gh/PaliC/22/orig -> origin/gh/PaliC/22/orig 2025-09-07T07:36:03.7026212Z * [new branch] gh/PaliC/23/base -> origin/gh/PaliC/23/base 2025-09-07T07:36:03.7027483Z * [new branch] gh/PaliC/23/head -> origin/gh/PaliC/23/head 2025-09-07T07:36:03.7028699Z * [new branch] gh/PaliC/23/orig -> origin/gh/PaliC/23/orig 2025-09-07T07:36:03.7030446Z * [new branch] gh/PaliC/24/base -> origin/gh/PaliC/24/base 2025-09-07T07:36:03.7031614Z * [new branch] gh/PaliC/24/head -> origin/gh/PaliC/24/head 2025-09-07T07:36:03.7032801Z * [new branch] gh/PaliC/24/orig -> origin/gh/PaliC/24/orig 2025-09-07T07:36:03.7034917Z * [new branch] gh/PaulZhang12/17/base -> origin/gh/PaulZhang12/17/base 2025-09-07T07:36:03.7036115Z * [new branch] gh/PaulZhang12/17/head -> origin/gh/PaulZhang12/17/head 2025-09-07T07:36:03.7038001Z * [new branch] gh/PaulZhang12/20/base -> origin/gh/PaulZhang12/20/base 2025-09-07T07:36:03.7039206Z * [new branch] gh/PaulZhang12/20/head -> origin/gh/PaulZhang12/20/head 2025-09-07T07:36:03.7040450Z * [new branch] gh/PaulZhang12/20/orig -> origin/gh/PaulZhang12/20/orig 2025-09-07T07:36:03.7042174Z * [new branch] gh/PaulZhang12/21/base -> origin/gh/PaulZhang12/21/base 2025-09-07T07:36:03.7043506Z * [new branch] gh/PaulZhang12/21/head -> origin/gh/PaulZhang12/21/head 2025-09-07T07:36:03.7044783Z * [new branch] gh/PaulZhang12/21/orig -> origin/gh/PaulZhang12/21/orig 2025-09-07T07:36:03.7046515Z * [new branch] gh/PaulZhang12/22/base -> origin/gh/PaulZhang12/22/base 2025-09-07T07:36:03.7047714Z * [new branch] gh/PaulZhang12/22/head -> origin/gh/PaulZhang12/22/head 2025-09-07T07:36:03.7048892Z * [new branch] gh/PaulZhang12/22/orig -> origin/gh/PaulZhang12/22/orig 2025-09-07T07:36:03.7050630Z * [new branch] gh/PaulZhang12/23/base -> origin/gh/PaulZhang12/23/base 2025-09-07T07:36:03.7051905Z * [new branch] gh/PaulZhang12/23/head -> origin/gh/PaulZhang12/23/head 2025-09-07T07:36:03.7053122Z * [new branch] gh/PaulZhang12/23/orig -> origin/gh/PaulZhang12/23/orig 2025-09-07T07:36:03.7054702Z * [new branch] gh/PaulZhang12/24/base -> origin/gh/PaulZhang12/24/base 2025-09-07T07:36:03.7055917Z * [new branch] gh/PaulZhang12/24/head -> origin/gh/PaulZhang12/24/head 2025-09-07T07:36:03.7057192Z * [new branch] gh/PaulZhang12/24/orig -> origin/gh/PaulZhang12/24/orig 2025-09-07T07:36:03.7058974Z * [new branch] gh/PaulZhang12/25/base -> origin/gh/PaulZhang12/25/base 2025-09-07T07:36:03.7060246Z * [new branch] gh/PaulZhang12/25/head -> origin/gh/PaulZhang12/25/head 2025-09-07T07:36:03.7061597Z * [new branch] gh/PaulZhang12/25/orig -> origin/gh/PaulZhang12/25/orig 2025-09-07T07:36:03.7063693Z * [new branch] gh/SamGinzburg/11/base -> origin/gh/SamGinzburg/11/base 2025-09-07T07:36:03.7065061Z * [new branch] gh/SamGinzburg/11/head -> origin/gh/SamGinzburg/11/head 2025-09-07T07:36:03.7067903Z * [new branch] gh/Sidharth123-cpu/24/base -> origin/gh/Sidharth123-cpu/24/base 2025-09-07T07:36:03.7069576Z * [new branch] gh/Sidharth123-cpu/25/base -> origin/gh/Sidharth123-cpu/25/base 2025-09-07T07:36:03.7071168Z * [new branch] gh/Sidharth123-cpu/26/base -> origin/gh/Sidharth123-cpu/26/base 2025-09-07T07:36:03.7072969Z * [new branch] gh/Sidharth123-cpu/27/base -> origin/gh/Sidharth123-cpu/27/base 2025-09-07T07:36:03.7075170Z * [new branch] gh/StrongerXi/1/base -> origin/gh/StrongerXi/1/base 2025-09-07T07:36:03.7076378Z * [new branch] gh/StrongerXi/1/head -> origin/gh/StrongerXi/1/head 2025-09-07T07:36:03.7078126Z * [new branch] gh/StrongerXi/133/base -> origin/gh/StrongerXi/133/base 2025-09-07T07:36:03.7079319Z * [new branch] gh/StrongerXi/133/head -> origin/gh/StrongerXi/133/head 2025-09-07T07:36:03.7080546Z * [new branch] gh/StrongerXi/133/orig -> origin/gh/StrongerXi/133/orig 2025-09-07T07:36:03.7082231Z * [new branch] gh/StrongerXi/134/base -> origin/gh/StrongerXi/134/base 2025-09-07T07:36:03.7083414Z * [new branch] gh/StrongerXi/134/head -> origin/gh/StrongerXi/134/head 2025-09-07T07:36:03.7084614Z * [new branch] gh/StrongerXi/134/orig -> origin/gh/StrongerXi/134/orig 2025-09-07T07:36:03.7086328Z * [new branch] gh/StrongerXi/136/base -> origin/gh/StrongerXi/136/base 2025-09-07T07:36:03.7087506Z * [new branch] gh/StrongerXi/136/head -> origin/gh/StrongerXi/136/head 2025-09-07T07:36:03.7088721Z * [new branch] gh/StrongerXi/136/orig -> origin/gh/StrongerXi/136/orig 2025-09-07T07:36:03.7090564Z * [new branch] gh/StrongerXi/137/base -> origin/gh/StrongerXi/137/base 2025-09-07T07:36:03.7091870Z * [new branch] gh/StrongerXi/137/head -> origin/gh/StrongerXi/137/head 2025-09-07T07:36:03.7093055Z * [new branch] gh/StrongerXi/137/orig -> origin/gh/StrongerXi/137/orig 2025-09-07T07:36:03.7094768Z * [new branch] gh/StrongerXi/138/base -> origin/gh/StrongerXi/138/base 2025-09-07T07:36:03.7096001Z * [new branch] gh/StrongerXi/138/head -> origin/gh/StrongerXi/138/head 2025-09-07T07:36:03.7097221Z * [new branch] gh/StrongerXi/138/orig -> origin/gh/StrongerXi/138/orig 2025-09-07T07:36:03.7099027Z * [new branch] gh/StrongerXi/139/base -> origin/gh/StrongerXi/139/base 2025-09-07T07:36:03.7100224Z * [new branch] gh/StrongerXi/139/head -> origin/gh/StrongerXi/139/head 2025-09-07T07:36:03.7101499Z * [new branch] gh/StrongerXi/139/orig -> origin/gh/StrongerXi/139/orig 2025-09-07T07:36:03.7103410Z * [new branch] gh/StrongerXi/140/base -> origin/gh/StrongerXi/140/base 2025-09-07T07:36:03.7104654Z * [new branch] gh/StrongerXi/140/head -> origin/gh/StrongerXi/140/head 2025-09-07T07:36:03.7105984Z * [new branch] gh/StrongerXi/140/orig -> origin/gh/StrongerXi/140/orig 2025-09-07T07:36:03.7107707Z * [new branch] gh/StrongerXi/71/base -> origin/gh/StrongerXi/71/base 2025-09-07T07:36:03.7108961Z * [new branch] gh/StrongerXi/71/head -> origin/gh/StrongerXi/71/head 2025-09-07T07:36:03.7110624Z * [new branch] gh/StrongerXi/72/base -> origin/gh/StrongerXi/72/base 2025-09-07T07:36:03.7111856Z * [new branch] gh/StrongerXi/72/head -> origin/gh/StrongerXi/72/head 2025-09-07T07:36:03.7114082Z * [new branch] gh/XilunWu/133/base -> origin/gh/XilunWu/133/base 2025-09-07T07:36:03.7115215Z * [new branch] gh/XilunWu/133/head -> origin/gh/XilunWu/133/head 2025-09-07T07:36:03.7116570Z * [new branch] gh/XilunWu/133/orig -> origin/gh/XilunWu/133/orig 2025-09-07T07:36:03.7118211Z * [new branch] gh/XilunWu/139/base -> origin/gh/XilunWu/139/base 2025-09-07T07:36:03.7119400Z * [new branch] gh/XilunWu/139/head -> origin/gh/XilunWu/139/head 2025-09-07T07:36:03.7120514Z * [new branch] gh/XilunWu/139/orig -> origin/gh/XilunWu/139/orig 2025-09-07T07:36:03.7122356Z * [new branch] gh/XilunWu/143/base -> origin/gh/XilunWu/143/base 2025-09-07T07:36:03.7123614Z * [new branch] gh/XilunWu/143/head -> origin/gh/XilunWu/143/head 2025-09-07T07:36:03.7124900Z * [new branch] gh/XilunWu/143/orig -> origin/gh/XilunWu/143/orig 2025-09-07T07:36:03.7126827Z * [new branch] gh/XilunWu/144/base -> origin/gh/XilunWu/144/base 2025-09-07T07:36:03.7128011Z * [new branch] gh/XilunWu/144/head -> origin/gh/XilunWu/144/head 2025-09-07T07:36:03.7129234Z * [new branch] gh/XilunWu/144/orig -> origin/gh/XilunWu/144/orig 2025-09-07T07:36:03.7130976Z * [new branch] gh/XilunWu/145/base -> origin/gh/XilunWu/145/base 2025-09-07T07:36:03.7132128Z * [new branch] gh/XilunWu/145/head -> origin/gh/XilunWu/145/head 2025-09-07T07:36:03.7133315Z * [new branch] gh/XilunWu/145/orig -> origin/gh/XilunWu/145/orig 2025-09-07T07:36:03.7134946Z * [new branch] gh/XilunWu/146/base -> origin/gh/XilunWu/146/base 2025-09-07T07:36:03.7136213Z * [new branch] gh/XilunWu/146/head -> origin/gh/XilunWu/146/head 2025-09-07T07:36:03.7137429Z * [new branch] gh/XilunWu/146/orig -> origin/gh/XilunWu/146/orig 2025-09-07T07:36:03.7139238Z * [new branch] gh/XilunWu/147/base -> origin/gh/XilunWu/147/base 2025-09-07T07:36:03.7140400Z * [new branch] gh/XilunWu/147/head -> origin/gh/XilunWu/147/head 2025-09-07T07:36:03.7141652Z * [new branch] gh/XilunWu/147/orig -> origin/gh/XilunWu/147/orig 2025-09-07T07:36:03.7143235Z * [new branch] gh/XilunWu/148/base -> origin/gh/XilunWu/148/base 2025-09-07T07:36:03.7144465Z * [new branch] gh/XilunWu/148/head -> origin/gh/XilunWu/148/head 2025-09-07T07:36:03.7145627Z * [new branch] gh/XilunWu/148/orig -> origin/gh/XilunWu/148/orig 2025-09-07T07:36:03.7147223Z * [new branch] gh/XilunWu/149/base -> origin/gh/XilunWu/149/base 2025-09-07T07:36:03.7148407Z * [new branch] gh/XilunWu/149/head -> origin/gh/XilunWu/149/head 2025-09-07T07:36:03.7149643Z * [new branch] gh/XilunWu/149/orig -> origin/gh/XilunWu/149/orig 2025-09-07T07:36:03.7151226Z * [new branch] gh/XilunWu/150/base -> origin/gh/XilunWu/150/base 2025-09-07T07:36:03.7152436Z * [new branch] gh/XilunWu/150/head -> origin/gh/XilunWu/150/head 2025-09-07T07:36:03.7153665Z * [new branch] gh/XilunWu/150/orig -> origin/gh/XilunWu/150/orig 2025-09-07T07:36:03.7155361Z * [new branch] gh/XilunWu/151/base -> origin/gh/XilunWu/151/base 2025-09-07T07:36:03.7156570Z * [new branch] gh/XilunWu/151/head -> origin/gh/XilunWu/151/head 2025-09-07T07:36:03.7157897Z * [new branch] gh/XilunWu/151/orig -> origin/gh/XilunWu/151/orig 2025-09-07T07:36:03.7159515Z * [new branch] gh/XilunWu/152/base -> origin/gh/XilunWu/152/base 2025-09-07T07:36:03.7160651Z * [new branch] gh/XilunWu/152/head -> origin/gh/XilunWu/152/head 2025-09-07T07:36:03.7161875Z * [new branch] gh/XilunWu/152/orig -> origin/gh/XilunWu/152/orig 2025-09-07T07:36:03.7163777Z * [new branch] gh/XilunWu/153/base -> origin/gh/XilunWu/153/base 2025-09-07T07:36:03.7165150Z * [new branch] gh/XilunWu/153/head -> origin/gh/XilunWu/153/head 2025-09-07T07:36:03.7166339Z * [new branch] gh/XilunWu/153/orig -> origin/gh/XilunWu/153/orig 2025-09-07T07:36:03.7168119Z * [new branch] gh/XilunWu/160/base -> origin/gh/XilunWu/160/base 2025-09-07T07:36:03.7169340Z * [new branch] gh/XilunWu/160/head -> origin/gh/XilunWu/160/head 2025-09-07T07:36:03.7170730Z * [new branch] gh/XilunWu/160/orig -> origin/gh/XilunWu/160/orig 2025-09-07T07:36:03.7172705Z * [new branch] gh/XilunWu/161/base -> origin/gh/XilunWu/161/base 2025-09-07T07:36:03.7173897Z * [new branch] gh/XilunWu/161/head -> origin/gh/XilunWu/161/head 2025-09-07T07:36:03.7175173Z * [new branch] gh/XilunWu/161/orig -> origin/gh/XilunWu/161/orig 2025-09-07T07:36:03.7177054Z * [new branch] gh/XilunWu/163/base -> origin/gh/XilunWu/163/base 2025-09-07T07:36:03.7178298Z * [new branch] gh/XilunWu/163/head -> origin/gh/XilunWu/163/head 2025-09-07T07:36:03.7179612Z * [new branch] gh/XilunWu/163/orig -> origin/gh/XilunWu/163/orig 2025-09-07T07:36:03.7182035Z * [new branch] gh/XilunWu/164/base -> origin/gh/XilunWu/164/base 2025-09-07T07:36:03.7183316Z * [new branch] gh/XilunWu/164/head -> origin/gh/XilunWu/164/head 2025-09-07T07:36:03.7184538Z * [new branch] gh/XilunWu/164/orig -> origin/gh/XilunWu/164/orig 2025-09-07T07:36:03.7186447Z * [new branch] gh/XilunWu/165/base -> origin/gh/XilunWu/165/base 2025-09-07T07:36:03.7187723Z * [new branch] gh/XilunWu/165/head -> origin/gh/XilunWu/165/head 2025-09-07T07:36:03.7188953Z * [new branch] gh/XilunWu/165/orig -> origin/gh/XilunWu/165/orig 2025-09-07T07:36:03.7190850Z * [new branch] gh/XilunWu/166/base -> origin/gh/XilunWu/166/base 2025-09-07T07:36:03.7192109Z * [new branch] gh/XilunWu/166/head -> origin/gh/XilunWu/166/head 2025-09-07T07:36:03.7193385Z * [new branch] gh/XilunWu/166/orig -> origin/gh/XilunWu/166/orig 2025-09-07T07:36:03.7195179Z * [new branch] gh/XilunWu/167/base -> origin/gh/XilunWu/167/base 2025-09-07T07:36:03.7196420Z * [new branch] gh/XilunWu/167/head -> origin/gh/XilunWu/167/head 2025-09-07T07:36:03.7197662Z * [new branch] gh/XilunWu/167/orig -> origin/gh/XilunWu/167/orig 2025-09-07T07:36:03.7199774Z * [new branch] gh/XilunWu/168/base -> origin/gh/XilunWu/168/base 2025-09-07T07:36:03.7200970Z * [new branch] gh/XilunWu/168/head -> origin/gh/XilunWu/168/head 2025-09-07T07:36:03.7202210Z * [new branch] gh/XilunWu/168/orig -> origin/gh/XilunWu/168/orig 2025-09-07T07:36:03.7204229Z * [new branch] gh/XilunWu/169/base -> origin/gh/XilunWu/169/base 2025-09-07T07:36:03.7205426Z * [new branch] gh/XilunWu/169/head -> origin/gh/XilunWu/169/head 2025-09-07T07:36:03.7206594Z * [new branch] gh/XilunWu/169/orig -> origin/gh/XilunWu/169/orig 2025-09-07T07:36:03.7208271Z * [new branch] gh/XilunWu/170/base -> origin/gh/XilunWu/170/base 2025-09-07T07:36:03.7209476Z * [new branch] gh/XilunWu/170/head -> origin/gh/XilunWu/170/head 2025-09-07T07:36:03.7210745Z * [new branch] gh/XilunWu/170/orig -> origin/gh/XilunWu/170/orig 2025-09-07T07:36:03.7212978Z * [new branch] gh/XuehaiPan/14/base -> origin/gh/XuehaiPan/14/base 2025-09-07T07:36:03.7214241Z * [new branch] gh/XuehaiPan/14/head -> origin/gh/XuehaiPan/14/head 2025-09-07T07:36:03.7215361Z * [new branch] gh/XuehaiPan/14/orig -> origin/gh/XuehaiPan/14/orig 2025-09-07T07:36:03.7217200Z * [new branch] gh/XuehaiPan/179/base -> origin/gh/XuehaiPan/179/base 2025-09-07T07:36:03.7218584Z * [new branch] gh/XuehaiPan/179/head -> origin/gh/XuehaiPan/179/head 2025-09-07T07:36:03.7219774Z * [new branch] gh/XuehaiPan/179/orig -> origin/gh/XuehaiPan/179/orig 2025-09-07T07:36:03.7221671Z * [new branch] gh/XuehaiPan/189/base -> origin/gh/XuehaiPan/189/base 2025-09-07T07:36:03.7222835Z * [new branch] gh/XuehaiPan/189/head -> origin/gh/XuehaiPan/189/head 2025-09-07T07:36:03.7224025Z * [new branch] gh/XuehaiPan/189/orig -> origin/gh/XuehaiPan/189/orig 2025-09-07T07:36:03.7225801Z * [new branch] gh/XuehaiPan/232/base -> origin/gh/XuehaiPan/232/base 2025-09-07T07:36:03.7227118Z * [new branch] gh/XuehaiPan/232/head -> origin/gh/XuehaiPan/232/head 2025-09-07T07:36:03.7228310Z * [new branch] gh/XuehaiPan/232/orig -> origin/gh/XuehaiPan/232/orig 2025-09-07T07:36:03.7230041Z * [new branch] gh/XuehaiPan/249/base -> origin/gh/XuehaiPan/249/base 2025-09-07T07:36:03.7231287Z * [new branch] gh/XuehaiPan/249/head -> origin/gh/XuehaiPan/249/head 2025-09-07T07:36:03.7232439Z * [new branch] gh/XuehaiPan/249/orig -> origin/gh/XuehaiPan/249/orig 2025-09-07T07:36:03.7234101Z * [new branch] gh/XuehaiPan/253/base -> origin/gh/XuehaiPan/253/base 2025-09-07T07:36:03.7235707Z * [new branch] gh/XuehaiPan/253/head -> origin/gh/XuehaiPan/253/head 2025-09-07T07:36:03.7236671Z * [new branch] gh/XuehaiPan/253/orig -> origin/gh/XuehaiPan/253/orig 2025-09-07T07:36:03.7238295Z * [new branch] gh/XuehaiPan/254/base -> origin/gh/XuehaiPan/254/base 2025-09-07T07:36:03.7239519Z * [new branch] gh/XuehaiPan/254/head -> origin/gh/XuehaiPan/254/head 2025-09-07T07:36:03.7240801Z * [new branch] gh/XuehaiPan/254/orig -> origin/gh/XuehaiPan/254/orig 2025-09-07T07:36:03.7242503Z * [new branch] gh/XuehaiPan/255/base -> origin/gh/XuehaiPan/255/base 2025-09-07T07:36:03.7243744Z * [new branch] gh/XuehaiPan/255/head -> origin/gh/XuehaiPan/255/head 2025-09-07T07:36:03.7244862Z * [new branch] gh/XuehaiPan/255/orig -> origin/gh/XuehaiPan/255/orig 2025-09-07T07:36:03.7246592Z * [new branch] gh/XuehaiPan/257/base -> origin/gh/XuehaiPan/257/base 2025-09-07T07:36:03.7247785Z * [new branch] gh/XuehaiPan/257/head -> origin/gh/XuehaiPan/257/head 2025-09-07T07:36:03.7248992Z * [new branch] gh/XuehaiPan/257/orig -> origin/gh/XuehaiPan/257/orig 2025-09-07T07:36:03.7250702Z * [new branch] gh/XuehaiPan/271/base -> origin/gh/XuehaiPan/271/base 2025-09-07T07:36:03.7251866Z * [new branch] gh/XuehaiPan/271/head -> origin/gh/XuehaiPan/271/head 2025-09-07T07:36:03.7253064Z * [new branch] gh/XuehaiPan/271/orig -> origin/gh/XuehaiPan/271/orig 2025-09-07T07:36:03.7254751Z * [new branch] gh/XuehaiPan/290/base -> origin/gh/XuehaiPan/290/base 2025-09-07T07:36:03.7256069Z * [new branch] gh/XuehaiPan/290/head -> origin/gh/XuehaiPan/290/head 2025-09-07T07:36:03.7257183Z * [new branch] gh/XuehaiPan/290/orig -> origin/gh/XuehaiPan/290/orig 2025-09-07T07:36:03.7259014Z * [new branch] gh/XuehaiPan/343/base -> origin/gh/XuehaiPan/343/base 2025-09-07T07:36:03.7266943Z * [new branch] gh/XuehaiPan/343/head -> origin/gh/XuehaiPan/343/head 2025-09-07T07:36:03.7267286Z * [new branch] gh/XuehaiPan/343/orig -> origin/gh/XuehaiPan/343/orig 2025-09-07T07:36:03.7267601Z * [new branch] gh/XuehaiPan/347/base -> origin/gh/XuehaiPan/347/base 2025-09-07T07:36:03.7267927Z * [new branch] gh/XuehaiPan/347/head -> origin/gh/XuehaiPan/347/head 2025-09-07T07:36:03.7268320Z * [new branch] gh/XuehaiPan/347/orig -> origin/gh/XuehaiPan/347/orig 2025-09-07T07:36:03.7268650Z * [new branch] gh/XuehaiPan/348/base -> origin/gh/XuehaiPan/348/base 2025-09-07T07:36:03.7268973Z * [new branch] gh/XuehaiPan/348/head -> origin/gh/XuehaiPan/348/head 2025-09-07T07:36:03.7270069Z * [new branch] gh/XuehaiPan/348/orig -> origin/gh/XuehaiPan/348/orig 2025-09-07T07:36:03.7271757Z * [new branch] gh/XuehaiPan/350/base -> origin/gh/XuehaiPan/350/base 2025-09-07T07:36:03.7272958Z * [new branch] gh/XuehaiPan/350/head -> origin/gh/XuehaiPan/350/head 2025-09-07T07:36:03.7274230Z * [new branch] gh/XuehaiPan/350/orig -> origin/gh/XuehaiPan/350/orig 2025-09-07T07:36:03.7276082Z * [new branch] gh/XuehaiPan/356/base -> origin/gh/XuehaiPan/356/base 2025-09-07T07:36:03.7277507Z * [new branch] gh/XuehaiPan/356/head -> origin/gh/XuehaiPan/356/head 2025-09-07T07:36:03.7278751Z * [new branch] gh/XuehaiPan/356/orig -> origin/gh/XuehaiPan/356/orig 2025-09-07T07:36:03.7280420Z * [new branch] gh/XuehaiPan/357/base -> origin/gh/XuehaiPan/357/base 2025-09-07T07:36:03.7281620Z * [new branch] gh/XuehaiPan/357/head -> origin/gh/XuehaiPan/357/head 2025-09-07T07:36:03.7282792Z * [new branch] gh/XuehaiPan/357/orig -> origin/gh/XuehaiPan/357/orig 2025-09-07T07:36:03.7284966Z * [new branch] gh/XuehaiPan/358/base -> origin/gh/XuehaiPan/358/base 2025-09-07T07:36:03.7286192Z * [new branch] gh/XuehaiPan/358/head -> origin/gh/XuehaiPan/358/head 2025-09-07T07:36:03.7287431Z * [new branch] gh/XuehaiPan/358/orig -> origin/gh/XuehaiPan/358/orig 2025-09-07T07:36:03.7289160Z * [new branch] gh/XuehaiPan/359/base -> origin/gh/XuehaiPan/359/base 2025-09-07T07:36:03.7290331Z * [new branch] gh/XuehaiPan/359/head -> origin/gh/XuehaiPan/359/head 2025-09-07T07:36:03.7291562Z * [new branch] gh/XuehaiPan/359/orig -> origin/gh/XuehaiPan/359/orig 2025-09-07T07:36:03.7293222Z * [new branch] gh/XuehaiPan/360/base -> origin/gh/XuehaiPan/360/base 2025-09-07T07:36:03.7294518Z * [new branch] gh/XuehaiPan/360/head -> origin/gh/XuehaiPan/360/head 2025-09-07T07:36:03.7295707Z * [new branch] gh/XuehaiPan/360/orig -> origin/gh/XuehaiPan/360/orig 2025-09-07T07:36:03.7297521Z * [new branch] gh/XuehaiPan/365/base -> origin/gh/XuehaiPan/365/base 2025-09-07T07:36:03.7298810Z * [new branch] gh/XuehaiPan/365/head -> origin/gh/XuehaiPan/365/head 2025-09-07T07:36:03.7299994Z * [new branch] gh/XuehaiPan/365/orig -> origin/gh/XuehaiPan/365/orig 2025-09-07T07:36:03.7301832Z * [new branch] gh/XuehaiPan/366/base -> origin/gh/XuehaiPan/366/base 2025-09-07T07:36:03.7303187Z * [new branch] gh/XuehaiPan/366/head -> origin/gh/XuehaiPan/366/head 2025-09-07T07:36:03.7305077Z * [new branch] gh/XuehaiPan/369/base -> origin/gh/XuehaiPan/369/base 2025-09-07T07:36:03.7306466Z * [new branch] gh/XuehaiPan/369/head -> origin/gh/XuehaiPan/369/head 2025-09-07T07:36:03.7307667Z * [new branch] gh/XuehaiPan/369/orig -> origin/gh/XuehaiPan/369/orig 2025-09-07T07:36:03.7309386Z * [new branch] gh/XuehaiPan/370/base -> origin/gh/XuehaiPan/370/base 2025-09-07T07:36:03.7310696Z * [new branch] gh/XuehaiPan/370/head -> origin/gh/XuehaiPan/370/head 2025-09-07T07:36:03.7311992Z * [new branch] gh/XuehaiPan/370/orig -> origin/gh/XuehaiPan/370/orig 2025-09-07T07:36:03.7313856Z * [new branch] gh/XuehaiPan/380/base -> origin/gh/XuehaiPan/380/base 2025-09-07T07:36:03.7315069Z * [new branch] gh/XuehaiPan/380/head -> origin/gh/XuehaiPan/380/head 2025-09-07T07:36:03.7316402Z * [new branch] gh/XuehaiPan/380/orig -> origin/gh/XuehaiPan/380/orig 2025-09-07T07:36:03.7318032Z * [new branch] gh/XuehaiPan/381/base -> origin/gh/XuehaiPan/381/base 2025-09-07T07:36:03.7319220Z * [new branch] gh/XuehaiPan/381/head -> origin/gh/XuehaiPan/381/head 2025-09-07T07:36:03.7320989Z * [new branch] gh/XuehaiPan/382/base -> origin/gh/XuehaiPan/382/base 2025-09-07T07:36:03.7322237Z * [new branch] gh/XuehaiPan/382/head -> origin/gh/XuehaiPan/382/head 2025-09-07T07:36:03.7323418Z * [new branch] gh/XuehaiPan/382/orig -> origin/gh/XuehaiPan/382/orig 2025-09-07T07:36:03.7325263Z * [new branch] gh/XuehaiPan/383/base -> origin/gh/XuehaiPan/383/base 2025-09-07T07:36:03.7326487Z * [new branch] gh/XuehaiPan/383/head -> origin/gh/XuehaiPan/383/head 2025-09-07T07:36:03.7327691Z * [new branch] gh/XuehaiPan/383/orig -> origin/gh/XuehaiPan/383/orig 2025-09-07T07:36:03.7329560Z * [new branch] gh/XuehaiPan/384/base -> origin/gh/XuehaiPan/384/base 2025-09-07T07:36:03.7330771Z * [new branch] gh/XuehaiPan/384/head -> origin/gh/XuehaiPan/384/head 2025-09-07T07:36:03.7331961Z * [new branch] gh/XuehaiPan/384/orig -> origin/gh/XuehaiPan/384/orig 2025-09-07T07:36:03.7333750Z * [new branch] gh/XuehaiPan/385/base -> origin/gh/XuehaiPan/385/base 2025-09-07T07:36:03.7334965Z * [new branch] gh/XuehaiPan/385/head -> origin/gh/XuehaiPan/385/head 2025-09-07T07:36:03.7336090Z * [new branch] gh/XuehaiPan/385/orig -> origin/gh/XuehaiPan/385/orig 2025-09-07T07:36:03.7337826Z * [new branch] gh/XuehaiPan/386/base -> origin/gh/XuehaiPan/386/base 2025-09-07T07:36:03.7339104Z * [new branch] gh/XuehaiPan/386/head -> origin/gh/XuehaiPan/386/head 2025-09-07T07:36:03.7340277Z * [new branch] gh/XuehaiPan/386/orig -> origin/gh/XuehaiPan/386/orig 2025-09-07T07:36:03.7342130Z * [new branch] gh/XuehaiPan/387/base -> origin/gh/XuehaiPan/387/base 2025-09-07T07:36:03.7343324Z * [new branch] gh/XuehaiPan/387/head -> origin/gh/XuehaiPan/387/head 2025-09-07T07:36:03.7344527Z * [new branch] gh/XuehaiPan/387/orig -> origin/gh/XuehaiPan/387/orig 2025-09-07T07:36:03.7346761Z * [new branch] gh/ZainRizvi/1/base -> origin/gh/ZainRizvi/1/base 2025-09-07T07:36:03.7347952Z * [new branch] gh/ZainRizvi/1/head -> origin/gh/ZainRizvi/1/head 2025-09-07T07:36:03.7349600Z * [new branch] gh/ZainRizvi/2/base -> origin/gh/ZainRizvi/2/base 2025-09-07T07:36:03.7350673Z * [new branch] gh/ZainRizvi/2/head -> origin/gh/ZainRizvi/2/head 2025-09-07T07:36:03.7352334Z * [new branch] gh/ZainRizvi/3/base -> origin/gh/ZainRizvi/3/base 2025-09-07T07:36:03.7353524Z * [new branch] gh/ZainRizvi/3/head -> origin/gh/ZainRizvi/3/head 2025-09-07T07:36:03.7355215Z * [new branch] gh/ZainRizvi/4/base -> origin/gh/ZainRizvi/4/base 2025-09-07T07:36:03.7356425Z * [new branch] gh/ZainRizvi/4/head -> origin/gh/ZainRizvi/4/head 2025-09-07T07:36:03.7358121Z * [new branch] gh/ZainRizvi/5/base -> origin/gh/ZainRizvi/5/base 2025-09-07T07:36:03.7359296Z * [new branch] gh/ZainRizvi/5/head -> origin/gh/ZainRizvi/5/head 2025-09-07T07:36:03.7361015Z * [new branch] gh/ZainRizvi/6/base -> origin/gh/ZainRizvi/6/base 2025-09-07T07:36:03.7362618Z * [new branch] gh/ZainRizvi/6/head -> origin/gh/ZainRizvi/6/head 2025-09-07T07:36:03.7363848Z * [new branch] gh/ZainRizvi/6/orig -> origin/gh/ZainRizvi/6/orig 2025-09-07T07:36:03.7365528Z * [new branch] gh/ZainRizvi/7/base -> origin/gh/ZainRizvi/7/base 2025-09-07T07:36:03.7366799Z * [new branch] gh/ZainRizvi/7/head -> origin/gh/ZainRizvi/7/head 2025-09-07T07:36:03.7367905Z * [new branch] gh/ZainRizvi/7/orig -> origin/gh/ZainRizvi/7/orig 2025-09-07T07:36:03.7369734Z * [new branch] gh/ZainRizvi/8/base -> origin/gh/ZainRizvi/8/base 2025-09-07T07:36:03.7370908Z * [new branch] gh/ZainRizvi/8/head -> origin/gh/ZainRizvi/8/head 2025-09-07T07:36:03.7372588Z * [new branch] gh/ZainRizvi/9/base -> origin/gh/ZainRizvi/9/base 2025-09-07T07:36:03.7373784Z * [new branch] gh/ZainRizvi/9/head -> origin/gh/ZainRizvi/9/head 2025-09-07T07:36:03.7374990Z * [new branch] gh/ZainRizvi/9/orig -> origin/gh/ZainRizvi/9/orig 2025-09-07T07:36:03.7377474Z * [new branch] gh/ZhiweiYan-96/39/base -> origin/gh/ZhiweiYan-96/39/base 2025-09-07T07:36:03.7378562Z * [new branch] gh/ZhiweiYan-96/39/head -> origin/gh/ZhiweiYan-96/39/head 2025-09-07T07:36:03.7379943Z * [new branch] gh/ZhiweiYan-96/39/orig -> origin/gh/ZhiweiYan-96/39/orig 2025-09-07T07:36:03.7381705Z * [new branch] gh/ZhiweiYan-96/44/base -> origin/gh/ZhiweiYan-96/44/base 2025-09-07T07:36:03.7382911Z * [new branch] gh/ZhiweiYan-96/44/head -> origin/gh/ZhiweiYan-96/44/head 2025-09-07T07:36:03.7384576Z * [new branch] gh/ZhiweiYan-96/45/base -> origin/gh/ZhiweiYan-96/45/base 2025-09-07T07:36:03.7385704Z * [new branch] gh/ZhiweiYan-96/45/head -> origin/gh/ZhiweiYan-96/45/head 2025-09-07T07:36:03.7387464Z * [new branch] gh/ZhiweiYan-96/49/base -> origin/gh/ZhiweiYan-96/49/base 2025-09-07T07:36:03.7388649Z * [new branch] gh/ZhiweiYan-96/49/head -> origin/gh/ZhiweiYan-96/49/head 2025-09-07T07:36:03.7390393Z * [new branch] gh/ZhiweiYan-96/62/base -> origin/gh/ZhiweiYan-96/62/base 2025-09-07T07:36:03.7391634Z * [new branch] gh/ZhiweiYan-96/62/head -> origin/gh/ZhiweiYan-96/62/head 2025-09-07T07:36:03.7393413Z * [new branch] gh/ZhiweiYan-96/64/base -> origin/gh/ZhiweiYan-96/64/base 2025-09-07T07:36:03.7394565Z * [new branch] gh/ZhiweiYan-96/64/head -> origin/gh/ZhiweiYan-96/64/head 2025-09-07T07:36:03.7395770Z * [new branch] gh/ZhiweiYan-96/64/orig -> origin/gh/ZhiweiYan-96/64/orig 2025-09-07T07:36:03.7397489Z * [new branch] gh/ZhiweiYan-96/65/base -> origin/gh/ZhiweiYan-96/65/base 2025-09-07T07:36:03.7398674Z * [new branch] gh/ZhiweiYan-96/65/head -> origin/gh/ZhiweiYan-96/65/head 2025-09-07T07:36:03.7399910Z * [new branch] gh/ZhiweiYan-96/65/orig -> origin/gh/ZhiweiYan-96/65/orig 2025-09-07T07:36:03.7401590Z * [new branch] gh/ZhiweiYan-96/66/base -> origin/gh/ZhiweiYan-96/66/base 2025-09-07T07:36:03.7402856Z * [new branch] gh/ZhiweiYan-96/66/head -> origin/gh/ZhiweiYan-96/66/head 2025-09-07T07:36:03.7404753Z * [new branch] gh/ZhiweiYan-96/67/base -> origin/gh/ZhiweiYan-96/67/base 2025-09-07T07:36:03.7405875Z * [new branch] gh/ZhiweiYan-96/67/head -> origin/gh/ZhiweiYan-96/67/head 2025-09-07T07:36:03.7407502Z * [new branch] gh/ZhiweiYan-96/68/base -> origin/gh/ZhiweiYan-96/68/base 2025-09-07T07:36:03.7408698Z * [new branch] gh/ZhiweiYan-96/68/head -> origin/gh/ZhiweiYan-96/68/head 2025-09-07T07:36:03.7409926Z * [new branch] gh/ZhiweiYan-96/68/orig -> origin/gh/ZhiweiYan-96/68/orig 2025-09-07T07:36:03.7412067Z * [new branch] gh/aakhundov/1/base -> origin/gh/aakhundov/1/base 2025-09-07T07:36:03.7413357Z * [new branch] gh/aakhundov/1/head -> origin/gh/aakhundov/1/head 2025-09-07T07:36:03.7414958Z * [new branch] gh/aakhundov/2/base -> origin/gh/aakhundov/2/base 2025-09-07T07:36:03.7416232Z * [new branch] gh/aakhundov/2/head -> origin/gh/aakhundov/2/head 2025-09-07T07:36:03.7418188Z * [new branch] gh/aditew01/openblas -> origin/gh/aditew01/openblas 2025-09-07T07:36:03.7419239Z * [new branch] gh/aditew01/sbgemm -> origin/gh/aditew01/sbgemm 2025-09-07T07:36:03.7420414Z * [new branch] gh/aditew01/vecbf16 -> origin/gh/aditew01/vecbf16 2025-09-07T07:36:03.7422331Z * [new branch] gh/alexbrauckmann/paddedtensor_faketensor_init -> origin/gh/alexbrauckmann/paddedtensor_faketensor_init 2025-09-07T07:36:03.7424412Z * [new branch] gh/alexsamardzic/9/base -> origin/gh/alexsamardzic/9/base 2025-09-07T07:36:03.7425581Z * [new branch] gh/alexsamardzic/9/head -> origin/gh/alexsamardzic/9/head 2025-09-07T07:36:03.7426813Z * [new branch] gh/alexsamardzic/9/orig -> origin/gh/alexsamardzic/9/orig 2025-09-07T07:36:03.7429016Z * [new branch] gh/amjames/18/base -> origin/gh/amjames/18/base 2025-09-07T07:36:03.7430161Z * [new branch] gh/amjames/18/head -> origin/gh/amjames/18/head 2025-09-07T07:36:03.7431352Z * [new branch] gh/amjames/18/orig -> origin/gh/amjames/18/orig 2025-09-07T07:36:03.7433672Z * [new branch] gh/andrewor14/35/base -> origin/gh/andrewor14/35/base 2025-09-07T07:36:03.7434939Z * [new branch] gh/andrewor14/35/head -> origin/gh/andrewor14/35/head 2025-09-07T07:36:03.7436142Z * [new branch] gh/andrewor14/35/orig -> origin/gh/andrewor14/35/orig 2025-09-07T07:36:03.7438008Z * [new branch] gh/andrewor14/50/base -> origin/gh/andrewor14/50/base 2025-09-07T07:36:03.7439391Z * [new branch] gh/andrewor14/50/head -> origin/gh/andrewor14/50/head 2025-09-07T07:36:03.7440641Z * [new branch] gh/andrewor14/50/orig -> origin/gh/andrewor14/50/orig 2025-09-07T07:36:03.7442355Z * [new branch] gh/andrewor14/51/base -> origin/gh/andrewor14/51/base 2025-09-07T07:36:03.7443624Z * [new branch] gh/andrewor14/51/orig -> origin/gh/andrewor14/51/orig 2025-09-07T07:36:03.7445824Z * [new branch] gh/andyanwang/1/base -> origin/gh/andyanwang/1/base 2025-09-07T07:36:03.7447005Z * [new branch] gh/andyanwang/1/head -> origin/gh/andyanwang/1/head 2025-09-07T07:36:03.7448162Z * [new branch] gh/andyanwang/1/orig -> origin/gh/andyanwang/1/orig 2025-09-07T07:36:03.7450083Z * [new branch] gh/andyanwang/13/base -> origin/gh/andyanwang/13/base 2025-09-07T07:36:03.7451302Z * [new branch] gh/andyanwang/13/head -> origin/gh/andyanwang/13/head 2025-09-07T07:36:03.7453034Z * [new branch] gh/andyanwang/13/orig -> origin/gh/andyanwang/13/orig 2025-09-07T07:36:03.7454877Z * [new branch] gh/andyanwang/2/base -> origin/gh/andyanwang/2/base 2025-09-07T07:36:03.7456221Z * [new branch] gh/andyanwang/2/head -> origin/gh/andyanwang/2/head 2025-09-07T07:36:03.7457438Z * [new branch] gh/andyanwang/2/orig -> origin/gh/andyanwang/2/orig 2025-09-07T07:36:03.7459277Z * [new branch] gh/andyanwang/28/base -> origin/gh/andyanwang/28/base 2025-09-07T07:36:03.7460509Z * [new branch] gh/andyanwang/28/head -> origin/gh/andyanwang/28/head 2025-09-07T07:36:03.7461734Z * [new branch] gh/andyanwang/28/orig -> origin/gh/andyanwang/28/orig 2025-09-07T07:36:03.7463325Z * [new branch] gh/andyanwang/3/base -> origin/gh/andyanwang/3/base 2025-09-07T07:36:03.7464550Z * [new branch] gh/andyanwang/3/head -> origin/gh/andyanwang/3/head 2025-09-07T07:36:03.7465766Z * [new branch] gh/andyanwang/3/orig -> origin/gh/andyanwang/3/orig 2025-09-07T07:36:03.7467569Z * [new branch] gh/andyanwang/30/base -> origin/gh/andyanwang/30/base 2025-09-07T07:36:03.7468972Z * [new branch] gh/andyanwang/30/orig -> origin/gh/andyanwang/30/orig 2025-09-07T07:36:03.7470629Z * [new branch] gh/andyanwang/31/base -> origin/gh/andyanwang/31/base 2025-09-07T07:36:03.7471966Z * [new branch] gh/andyanwang/31/orig -> origin/gh/andyanwang/31/orig 2025-09-07T07:36:03.7474042Z * [new branch] gh/andyanwang/32/base -> origin/gh/andyanwang/32/base 2025-09-07T07:36:03.7475228Z * [new branch] gh/andyanwang/32/head -> origin/gh/andyanwang/32/head 2025-09-07T07:36:03.7476542Z * [new branch] gh/andyanwang/32/orig -> origin/gh/andyanwang/32/orig 2025-09-07T07:36:03.7478274Z * [new branch] gh/andyanwang/39/base -> origin/gh/andyanwang/39/base 2025-09-07T07:36:03.7479499Z * [new branch] gh/andyanwang/39/head -> origin/gh/andyanwang/39/head 2025-09-07T07:36:03.7480737Z * [new branch] gh/andyanwang/39/orig -> origin/gh/andyanwang/39/orig 2025-09-07T07:36:03.7482544Z * [new branch] gh/andyanwang/4/base -> origin/gh/andyanwang/4/base 2025-09-07T07:36:03.7483679Z * [new branch] gh/andyanwang/4/head -> origin/gh/andyanwang/4/head 2025-09-07T07:36:03.7484919Z * [new branch] gh/andyanwang/4/orig -> origin/gh/andyanwang/4/orig 2025-09-07T07:36:03.7487009Z * [new branch] gh/angelayi/107/base -> origin/gh/angelayi/107/base 2025-09-07T07:36:03.7488290Z * [new branch] gh/angelayi/107/head -> origin/gh/angelayi/107/head 2025-09-07T07:36:03.7490051Z * [new branch] gh/angelayi/111/base -> origin/gh/angelayi/111/base 2025-09-07T07:36:03.7491378Z * [new branch] gh/angelayi/111/head -> origin/gh/angelayi/111/head 2025-09-07T07:36:03.7492557Z * [new branch] gh/angelayi/111/orig -> origin/gh/angelayi/111/orig 2025-09-07T07:36:03.7494269Z * [new branch] gh/angelayi/112/base -> origin/gh/angelayi/112/base 2025-09-07T07:36:03.7495646Z * [new branch] gh/angelayi/112/head -> origin/gh/angelayi/112/head 2025-09-07T07:36:03.7496878Z * [new branch] gh/angelayi/112/orig -> origin/gh/angelayi/112/orig 2025-09-07T07:36:03.7498801Z * [new branch] gh/angelayi/113/base -> origin/gh/angelayi/113/base 2025-09-07T07:36:03.7499914Z * [new branch] gh/angelayi/113/head -> origin/gh/angelayi/113/head 2025-09-07T07:36:03.7501062Z * [new branch] gh/angelayi/113/orig -> origin/gh/angelayi/113/orig 2025-09-07T07:36:03.7502800Z * [new branch] gh/angelayi/114/base -> origin/gh/angelayi/114/base 2025-09-07T07:36:03.7504186Z * [new branch] gh/angelayi/114/head -> origin/gh/angelayi/114/head 2025-09-07T07:36:03.7505430Z * [new branch] gh/angelayi/114/orig -> origin/gh/angelayi/114/orig 2025-09-07T07:36:03.7507210Z * [new branch] gh/angelayi/115/base -> origin/gh/angelayi/115/base 2025-09-07T07:36:03.7508452Z * [new branch] gh/angelayi/115/head -> origin/gh/angelayi/115/head 2025-09-07T07:36:03.7509621Z * [new branch] gh/angelayi/115/orig -> origin/gh/angelayi/115/orig 2025-09-07T07:36:03.7511827Z * [new branch] gh/anijain2305/753/base -> origin/gh/anijain2305/753/base 2025-09-07T07:36:03.7513021Z * [new branch] gh/anijain2305/753/head -> origin/gh/anijain2305/753/head 2025-09-07T07:36:03.7514205Z * [new branch] gh/anijain2305/753/orig -> origin/gh/anijain2305/753/orig 2025-09-07T07:36:03.7516070Z * [new branch] gh/anijain2305/766/base -> origin/gh/anijain2305/766/base 2025-09-07T07:36:03.7517174Z * [new branch] gh/anijain2305/766/head -> origin/gh/anijain2305/766/head 2025-09-07T07:36:03.7518495Z * [new branch] gh/anijain2305/766/orig -> origin/gh/anijain2305/766/orig 2025-09-07T07:36:03.7520206Z * [new branch] gh/anijain2305/790/base -> origin/gh/anijain2305/790/base 2025-09-07T07:36:03.7521286Z * [new branch] gh/anijain2305/790/head -> origin/gh/anijain2305/790/head 2025-09-07T07:36:03.7522582Z * [new branch] gh/anijain2305/790/orig -> origin/gh/anijain2305/790/orig 2025-09-07T07:36:03.7524246Z * [new branch] gh/anijain2305/792/base -> origin/gh/anijain2305/792/base 2025-09-07T07:36:03.7525473Z * [new branch] gh/anijain2305/792/head -> origin/gh/anijain2305/792/head 2025-09-07T07:36:03.7526843Z * [new branch] gh/anijain2305/792/orig -> origin/gh/anijain2305/792/orig 2025-09-07T07:36:03.7528515Z * [new branch] gh/anijain2305/803/base -> origin/gh/anijain2305/803/base 2025-09-07T07:36:03.7529725Z * [new branch] gh/anijain2305/803/head -> origin/gh/anijain2305/803/head 2025-09-07T07:36:03.7530900Z * [new branch] gh/anijain2305/803/orig -> origin/gh/anijain2305/803/orig 2025-09-07T07:36:03.7532596Z * [new branch] gh/anijain2305/804/base -> origin/gh/anijain2305/804/base 2025-09-07T07:36:03.7533759Z * [new branch] gh/anijain2305/804/head -> origin/gh/anijain2305/804/head 2025-09-07T07:36:03.7534992Z * [new branch] gh/anijain2305/804/orig -> origin/gh/anijain2305/804/orig 2025-09-07T07:36:03.7536683Z * [new branch] gh/anijain2305/805/base -> origin/gh/anijain2305/805/base 2025-09-07T07:36:03.7537944Z * [new branch] gh/anijain2305/805/head -> origin/gh/anijain2305/805/head 2025-09-07T07:36:03.7539212Z * [new branch] gh/anijain2305/805/orig -> origin/gh/anijain2305/805/orig 2025-09-07T07:36:03.7541173Z * [new branch] gh/anijain2305/810/base -> origin/gh/anijain2305/810/base 2025-09-07T07:36:03.7542511Z * [new branch] gh/anijain2305/810/head -> origin/gh/anijain2305/810/head 2025-09-07T07:36:03.7543503Z * [new branch] gh/anijain2305/810/orig -> origin/gh/anijain2305/810/orig 2025-09-07T07:36:03.7545297Z * [new branch] gh/anijain2305/812/base -> origin/gh/anijain2305/812/base 2025-09-07T07:36:03.7546537Z * [new branch] gh/anijain2305/812/head -> origin/gh/anijain2305/812/head 2025-09-07T07:36:03.7547709Z * [new branch] gh/anijain2305/812/orig -> origin/gh/anijain2305/812/orig 2025-09-07T07:36:03.7549461Z * [new branch] gh/anijain2305/838/base -> origin/gh/anijain2305/838/base 2025-09-07T07:36:03.7550657Z * [new branch] gh/anijain2305/838/head -> origin/gh/anijain2305/838/head 2025-09-07T07:36:03.7551871Z * [new branch] gh/anijain2305/838/orig -> origin/gh/anijain2305/838/orig 2025-09-07T07:36:03.7553554Z * [new branch] gh/anijain2305/839/base -> origin/gh/anijain2305/839/base 2025-09-07T07:36:03.7554827Z * [new branch] gh/anijain2305/839/head -> origin/gh/anijain2305/839/head 2025-09-07T07:36:03.7556000Z * [new branch] gh/anijain2305/839/orig -> origin/gh/anijain2305/839/orig 2025-09-07T07:36:03.7557688Z * [new branch] gh/anijain2305/843/base -> origin/gh/anijain2305/843/base 2025-09-07T07:36:03.7558867Z * [new branch] gh/anijain2305/843/head -> origin/gh/anijain2305/843/head 2025-09-07T07:36:03.7560071Z * [new branch] gh/anijain2305/843/orig -> origin/gh/anijain2305/843/orig 2025-09-07T07:36:03.7561949Z * [new branch] gh/anijain2305/844/base -> origin/gh/anijain2305/844/base 2025-09-07T07:36:03.7563130Z * [new branch] gh/anijain2305/844/head -> origin/gh/anijain2305/844/head 2025-09-07T07:36:03.7564352Z * [new branch] gh/anijain2305/844/orig -> origin/gh/anijain2305/844/orig 2025-09-07T07:36:03.7566096Z * [new branch] gh/anijain2305/846/base -> origin/gh/anijain2305/846/base 2025-09-07T07:36:03.7567381Z * [new branch] gh/anijain2305/846/head -> origin/gh/anijain2305/846/head 2025-09-07T07:36:03.7568514Z * [new branch] gh/anijain2305/846/orig -> origin/gh/anijain2305/846/orig 2025-09-07T07:36:03.7570278Z * [new branch] gh/anijain2305/848/base -> origin/gh/anijain2305/848/base 2025-09-07T07:36:03.7571552Z * [new branch] gh/anijain2305/848/head -> origin/gh/anijain2305/848/head 2025-09-07T07:36:03.7572880Z * [new branch] gh/anijain2305/848/orig -> origin/gh/anijain2305/848/orig 2025-09-07T07:36:03.7574599Z * [new branch] gh/anijain2305/849/base -> origin/gh/anijain2305/849/base 2025-09-07T07:36:03.7575825Z * [new branch] gh/anijain2305/849/head -> origin/gh/anijain2305/849/head 2025-09-07T07:36:03.7576996Z * [new branch] gh/anijain2305/849/orig -> origin/gh/anijain2305/849/orig 2025-09-07T07:36:03.7579451Z * [new branch] gh/anijain2305/850/base -> origin/gh/anijain2305/850/base 2025-09-07T07:36:03.7580349Z * [new branch] gh/anijain2305/850/head -> origin/gh/anijain2305/850/head 2025-09-07T07:36:03.7581584Z * [new branch] gh/anijain2305/850/orig -> origin/gh/anijain2305/850/orig 2025-09-07T07:36:03.7583393Z * [new branch] gh/anijain2305/851/base -> origin/gh/anijain2305/851/base 2025-09-07T07:36:03.7584645Z * [new branch] gh/anijain2305/851/head -> origin/gh/anijain2305/851/head 2025-09-07T07:36:03.7585808Z * [new branch] gh/anijain2305/851/orig -> origin/gh/anijain2305/851/orig 2025-09-07T07:36:03.7587670Z * [new branch] gh/anijain2305/852/base -> origin/gh/anijain2305/852/base 2025-09-07T07:36:03.7588908Z * [new branch] gh/anijain2305/852/head -> origin/gh/anijain2305/852/head 2025-09-07T07:36:03.7590184Z * [new branch] gh/anijain2305/852/orig -> origin/gh/anijain2305/852/orig 2025-09-07T07:36:03.7591931Z * [new branch] gh/anijain2305/853/base -> origin/gh/anijain2305/853/base 2025-09-07T07:36:03.7593054Z * [new branch] gh/anijain2305/853/head -> origin/gh/anijain2305/853/head 2025-09-07T07:36:03.7594209Z * [new branch] gh/anijain2305/853/orig -> origin/gh/anijain2305/853/orig 2025-09-07T07:36:03.7595997Z * [new branch] gh/anijain2305/854/base -> origin/gh/anijain2305/854/base 2025-09-07T07:36:03.7597369Z * [new branch] gh/anijain2305/854/head -> origin/gh/anijain2305/854/head 2025-09-07T07:36:03.7598549Z * [new branch] gh/anijain2305/854/orig -> origin/gh/anijain2305/854/orig 2025-09-07T07:36:03.7600376Z * [new branch] gh/anijain2305/855/base -> origin/gh/anijain2305/855/base 2025-09-07T07:36:03.7601559Z * [new branch] gh/anijain2305/855/head -> origin/gh/anijain2305/855/head 2025-09-07T07:36:03.7602798Z * [new branch] gh/anijain2305/855/orig -> origin/gh/anijain2305/855/orig 2025-09-07T07:36:03.7606660Z * [new branch] gh/anijain2305/856/base -> origin/gh/anijain2305/856/base 2025-09-07T07:36:03.7607869Z * [new branch] gh/anijain2305/856/head -> origin/gh/anijain2305/856/head 2025-09-07T07:36:03.7609123Z * [new branch] gh/anijain2305/856/orig -> origin/gh/anijain2305/856/orig 2025-09-07T07:36:03.7610871Z * [new branch] gh/anijain2305/857/base -> origin/gh/anijain2305/857/base 2025-09-07T07:36:03.7612075Z * [new branch] gh/anijain2305/857/head -> origin/gh/anijain2305/857/head 2025-09-07T07:36:03.7613274Z * [new branch] gh/anijain2305/857/orig -> origin/gh/anijain2305/857/orig 2025-09-07T07:36:03.7614993Z * [new branch] gh/anijain2305/858/base -> origin/gh/anijain2305/858/base 2025-09-07T07:36:03.7616179Z * [new branch] gh/anijain2305/858/head -> origin/gh/anijain2305/858/head 2025-09-07T07:36:03.7617472Z * [new branch] gh/anijain2305/858/orig -> origin/gh/anijain2305/858/orig 2025-09-07T07:36:03.7619288Z * [new branch] gh/anijain2305/859/base -> origin/gh/anijain2305/859/base 2025-09-07T07:36:03.7620490Z * [new branch] gh/anijain2305/859/head -> origin/gh/anijain2305/859/head 2025-09-07T07:36:03.7621684Z * [new branch] gh/anijain2305/859/orig -> origin/gh/anijain2305/859/orig 2025-09-07T07:36:03.7623427Z * [new branch] gh/anijain2305/860/base -> origin/gh/anijain2305/860/base 2025-09-07T07:36:03.7624651Z * [new branch] gh/anijain2305/860/head -> origin/gh/anijain2305/860/head 2025-09-07T07:36:03.7625905Z * [new branch] gh/anijain2305/860/orig -> origin/gh/anijain2305/860/orig 2025-09-07T07:36:03.7627679Z * [new branch] gh/anijain2305/861/base -> origin/gh/anijain2305/861/base 2025-09-07T07:36:03.7628858Z * [new branch] gh/anijain2305/861/head -> origin/gh/anijain2305/861/head 2025-09-07T07:36:03.7630129Z * [new branch] gh/anijain2305/861/orig -> origin/gh/anijain2305/861/orig 2025-09-07T07:36:03.7631913Z * [new branch] gh/anijain2305/862/base -> origin/gh/anijain2305/862/base 2025-09-07T07:36:03.7633181Z * [new branch] gh/anijain2305/862/head -> origin/gh/anijain2305/862/head 2025-09-07T07:36:03.7634426Z * [new branch] gh/anijain2305/862/orig -> origin/gh/anijain2305/862/orig 2025-09-07T07:36:03.7636324Z * [new branch] gh/anijain2305/863/base -> origin/gh/anijain2305/863/base 2025-09-07T07:36:03.7637560Z * [new branch] gh/anijain2305/863/head -> origin/gh/anijain2305/863/head 2025-09-07T07:36:03.7638793Z * [new branch] gh/anijain2305/863/orig -> origin/gh/anijain2305/863/orig 2025-09-07T07:36:03.7640605Z * [new branch] gh/anijain2305/864/base -> origin/gh/anijain2305/864/base 2025-09-07T07:36:03.7641790Z * [new branch] gh/anijain2305/864/head -> origin/gh/anijain2305/864/head 2025-09-07T07:36:03.7643029Z * [new branch] gh/anijain2305/864/orig -> origin/gh/anijain2305/864/orig 2025-09-07T07:36:03.7644809Z * [new branch] gh/anijain2305/865/base -> origin/gh/anijain2305/865/base 2025-09-07T07:36:03.7646003Z * [new branch] gh/anijain2305/865/head -> origin/gh/anijain2305/865/head 2025-09-07T07:36:03.7647155Z * [new branch] gh/anijain2305/865/orig -> origin/gh/anijain2305/865/orig 2025-09-07T07:36:03.7648932Z * [new branch] gh/anijain2305/866/base -> origin/gh/anijain2305/866/base 2025-09-07T07:36:03.7650123Z * [new branch] gh/anijain2305/866/head -> origin/gh/anijain2305/866/head 2025-09-07T07:36:03.7651322Z * [new branch] gh/anijain2305/866/orig -> origin/gh/anijain2305/866/orig 2025-09-07T07:36:03.7653584Z * [new branch] gh/anjali411/216/base -> origin/gh/anjali411/216/base 2025-09-07T07:36:03.7654739Z * [new branch] gh/anjali411/216/head -> origin/gh/anjali411/216/head 2025-09-07T07:36:03.7655928Z * [new branch] gh/anjali411/216/orig -> origin/gh/anjali411/216/orig 2025-09-07T07:36:03.7658286Z * [new branch] gh/ankitageorge/13/base -> origin/gh/ankitageorge/13/base 2025-09-07T07:36:03.7659512Z * [new branch] gh/ankitageorge/13/head -> origin/gh/ankitageorge/13/head 2025-09-07T07:36:03.7660756Z * [new branch] gh/ankitageorge/13/orig -> origin/gh/ankitageorge/13/orig 2025-09-07T07:36:03.7662585Z * [new branch] gh/ankitageorge/14/base -> origin/gh/ankitageorge/14/base 2025-09-07T07:36:03.7663717Z * [new branch] gh/ankitageorge/14/head -> origin/gh/ankitageorge/14/head 2025-09-07T07:36:03.7665077Z * [new branch] gh/ankitageorge/14/orig -> origin/gh/ankitageorge/14/orig 2025-09-07T07:36:03.7666922Z * [new branch] gh/ankitageorge/15/base -> origin/gh/ankitageorge/15/base 2025-09-07T07:36:03.7668114Z * [new branch] gh/ankitageorge/15/head -> origin/gh/ankitageorge/15/head 2025-09-07T07:36:03.7669377Z * [new branch] gh/ankitageorge/15/orig -> origin/gh/ankitageorge/15/orig 2025-09-07T07:36:03.7671177Z * [new branch] gh/ankitageorge/16/base -> origin/gh/ankitageorge/16/base 2025-09-07T07:36:03.7672554Z * [new branch] gh/ankitageorge/16/head -> origin/gh/ankitageorge/16/head 2025-09-07T07:36:03.7673822Z * [new branch] gh/ankitageorge/16/orig -> origin/gh/ankitageorge/16/orig 2025-09-07T07:36:03.7675836Z * [new branch] gh/ankitageorge/17/base -> origin/gh/ankitageorge/17/base 2025-09-07T07:36:03.7677020Z * [new branch] gh/ankitageorge/17/head -> origin/gh/ankitageorge/17/head 2025-09-07T07:36:03.7678238Z * [new branch] gh/ankitageorge/17/orig -> origin/gh/ankitageorge/17/orig 2025-09-07T07:36:03.7680098Z * [new branch] gh/ankitageorge/21/base -> origin/gh/ankitageorge/21/base 2025-09-07T07:36:03.7681224Z * [new branch] gh/ankitageorge/21/head -> origin/gh/ankitageorge/21/head 2025-09-07T07:36:03.7682406Z * [new branch] gh/ankitageorge/21/orig -> origin/gh/ankitageorge/21/orig 2025-09-07T07:36:03.7684712Z * [new branch] gh/anshul-si/1/base -> origin/gh/anshul-si/1/base 2025-09-07T07:36:03.7685911Z * [new branch] gh/anshul-si/1/head -> origin/gh/anshul-si/1/head 2025-09-07T07:36:03.7687626Z * [new branch] gh/anshul-si/15/base -> origin/gh/anshul-si/15/base 2025-09-07T07:36:03.7688831Z * [new branch] gh/anshul-si/15/head -> origin/gh/anshul-si/15/head 2025-09-07T07:36:03.7690019Z * [new branch] gh/anshul-si/15/orig -> origin/gh/anshul-si/15/orig 2025-09-07T07:36:03.7691943Z * [new branch] gh/anshul-si/16/base -> origin/gh/anshul-si/16/base 2025-09-07T07:36:03.7693101Z * [new branch] gh/anshul-si/16/head -> origin/gh/anshul-si/16/head 2025-09-07T07:36:03.7694334Z * [new branch] gh/anshul-si/16/orig -> origin/gh/anshul-si/16/orig 2025-09-07T07:36:03.7696192Z * [new branch] gh/anshul-si/17/base -> origin/gh/anshul-si/17/base 2025-09-07T07:36:03.7697450Z * [new branch] gh/anshul-si/17/head -> origin/gh/anshul-si/17/head 2025-09-07T07:36:03.7699101Z * [new branch] gh/anshul-si/17/orig -> origin/gh/anshul-si/17/orig 2025-09-07T07:36:03.7700772Z * [new branch] gh/anshul-si/18/base -> origin/gh/anshul-si/18/base 2025-09-07T07:36:03.7701952Z * [new branch] gh/anshul-si/18/head -> origin/gh/anshul-si/18/head 2025-09-07T07:36:03.7703400Z * [new branch] gh/anshul-si/18/orig -> origin/gh/anshul-si/18/orig 2025-09-07T07:36:03.7705384Z * [new branch] gh/anshul-si/19/base -> origin/gh/anshul-si/19/base 2025-09-07T07:36:03.7706457Z * [new branch] gh/anshul-si/19/head -> origin/gh/anshul-si/19/head 2025-09-07T07:36:03.7707692Z * [new branch] gh/anshul-si/19/orig -> origin/gh/anshul-si/19/orig 2025-09-07T07:36:03.7709511Z * [new branch] gh/anshul-si/2/base -> origin/gh/anshul-si/2/base 2025-09-07T07:36:03.7710680Z * [new branch] gh/anshul-si/2/head -> origin/gh/anshul-si/2/head 2025-09-07T07:36:03.7712607Z * [new branch] gh/anshul-si/20/base -> origin/gh/anshul-si/20/base 2025-09-07T07:36:03.7713815Z * [new branch] gh/anshul-si/20/head -> origin/gh/anshul-si/20/head 2025-09-07T07:36:03.7714995Z * [new branch] gh/anshul-si/20/orig -> origin/gh/anshul-si/20/orig 2025-09-07T07:36:03.7716677Z * [new branch] gh/anshul-si/21/base -> origin/gh/anshul-si/21/base 2025-09-07T07:36:03.7718026Z * [new branch] gh/anshul-si/21/head -> origin/gh/anshul-si/21/head 2025-09-07T07:36:03.7719126Z * [new branch] gh/anshul-si/21/orig -> origin/gh/anshul-si/21/orig 2025-09-07T07:36:03.7720943Z * [new branch] gh/anshul-si/22/base -> origin/gh/anshul-si/22/base 2025-09-07T07:36:03.7722177Z * [new branch] gh/anshul-si/22/head -> origin/gh/anshul-si/22/head 2025-09-07T07:36:03.7723392Z * [new branch] gh/anshul-si/22/orig -> origin/gh/anshul-si/22/orig 2025-09-07T07:36:03.7725028Z * [new branch] gh/anshul-si/23/base -> origin/gh/anshul-si/23/base 2025-09-07T07:36:03.7726295Z * [new branch] gh/anshul-si/23/head -> origin/gh/anshul-si/23/head 2025-09-07T07:36:03.7727506Z * [new branch] gh/anshul-si/23/orig -> origin/gh/anshul-si/23/orig 2025-09-07T07:36:03.7729237Z * [new branch] gh/anshul-si/24/base -> origin/gh/anshul-si/24/base 2025-09-07T07:36:03.7730501Z * [new branch] gh/anshul-si/24/head -> origin/gh/anshul-si/24/head 2025-09-07T07:36:03.7731691Z * [new branch] gh/anshul-si/24/orig -> origin/gh/anshul-si/24/orig 2025-09-07T07:36:03.7733450Z * [new branch] gh/anshul-si/25/base -> origin/gh/anshul-si/25/base 2025-09-07T07:36:03.7734678Z * [new branch] gh/anshul-si/25/head -> origin/gh/anshul-si/25/head 2025-09-07T07:36:03.7735875Z * [new branch] gh/anshul-si/25/orig -> origin/gh/anshul-si/25/orig 2025-09-07T07:36:03.7737655Z * [new branch] gh/anshul-si/26/base -> origin/gh/anshul-si/26/base 2025-09-07T07:36:03.7738893Z * [new branch] gh/anshul-si/26/head -> origin/gh/anshul-si/26/head 2025-09-07T07:36:03.7740061Z * [new branch] gh/anshul-si/26/orig -> origin/gh/anshul-si/26/orig 2025-09-07T07:36:03.7741819Z * [new branch] gh/anshul-si/27/base -> origin/gh/anshul-si/27/base 2025-09-07T07:36:03.7743095Z * [new branch] gh/anshul-si/27/head -> origin/gh/anshul-si/27/head 2025-09-07T07:36:03.7744842Z * [new branch] gh/anshul-si/27/orig -> origin/gh/anshul-si/27/orig 2025-09-07T07:36:03.7746483Z * [new branch] gh/anshul-si/28/base -> origin/gh/anshul-si/28/base 2025-09-07T07:36:03.7747662Z * [new branch] gh/anshul-si/28/head -> origin/gh/anshul-si/28/head 2025-09-07T07:36:03.7748861Z * [new branch] gh/anshul-si/28/orig -> origin/gh/anshul-si/28/orig 2025-09-07T07:36:03.7750456Z * [new branch] gh/anshul-si/29/base -> origin/gh/anshul-si/29/base 2025-09-07T07:36:03.7751807Z * [new branch] gh/anshul-si/29/head -> origin/gh/anshul-si/29/head 2025-09-07T07:36:03.7753006Z * [new branch] gh/anshul-si/29/orig -> origin/gh/anshul-si/29/orig 2025-09-07T07:36:03.7754670Z * [new branch] gh/anshul-si/3/base -> origin/gh/anshul-si/3/base 2025-09-07T07:36:03.7755793Z * [new branch] gh/anshul-si/3/head -> origin/gh/anshul-si/3/head 2025-09-07T07:36:03.7757432Z * [new branch] gh/anshul-si/4/base -> origin/gh/anshul-si/4/base 2025-09-07T07:36:03.7758536Z * [new branch] gh/anshul-si/4/head -> origin/gh/anshul-si/4/head 2025-09-07T07:36:03.7760598Z * [new branch] gh/anshul-si/5/base -> origin/gh/anshul-si/5/base 2025-09-07T07:36:03.7761813Z * [new branch] gh/anshul-si/5/head -> origin/gh/anshul-si/5/head 2025-09-07T07:36:03.7764157Z * [new branch] gh/aorenste/132/base -> origin/gh/aorenste/132/base 2025-09-07T07:36:03.7765450Z * [new branch] gh/aorenste/132/head -> origin/gh/aorenste/132/head 2025-09-07T07:36:03.7767637Z * [new branch] gh/bdhirsh/650/base -> origin/gh/bdhirsh/650/base 2025-09-07T07:36:03.7769047Z * [new branch] gh/bdhirsh/650/head -> origin/gh/bdhirsh/650/head 2025-09-07T07:36:03.7770178Z * [new branch] gh/bdhirsh/650/orig -> origin/gh/bdhirsh/650/orig 2025-09-07T07:36:03.7771935Z * [new branch] gh/bdhirsh/663/base -> origin/gh/bdhirsh/663/base 2025-09-07T07:36:03.7773174Z * [new branch] gh/bdhirsh/663/head -> origin/gh/bdhirsh/663/head 2025-09-07T07:36:03.7774351Z * [new branch] gh/bdhirsh/663/orig -> origin/gh/bdhirsh/663/orig 2025-09-07T07:36:03.7776231Z * [new branch] gh/bdhirsh/665/base -> origin/gh/bdhirsh/665/base 2025-09-07T07:36:03.7777434Z * [new branch] gh/bdhirsh/665/head -> origin/gh/bdhirsh/665/head 2025-09-07T07:36:03.7778752Z * [new branch] gh/bdhirsh/665/orig -> origin/gh/bdhirsh/665/orig 2025-09-07T07:36:03.7780696Z * [new branch] gh/bdhirsh/666/base -> origin/gh/bdhirsh/666/base 2025-09-07T07:36:03.7782121Z * [new branch] gh/bdhirsh/666/head -> origin/gh/bdhirsh/666/head 2025-09-07T07:36:03.7783340Z * [new branch] gh/bdhirsh/666/orig -> origin/gh/bdhirsh/666/orig 2025-09-07T07:36:03.7785312Z * [new branch] gh/bdhirsh/667/base -> origin/gh/bdhirsh/667/base 2025-09-07T07:36:03.7786464Z * [new branch] gh/bdhirsh/667/head -> origin/gh/bdhirsh/667/head 2025-09-07T07:36:03.7787720Z * [new branch] gh/bdhirsh/667/orig -> origin/gh/bdhirsh/667/orig 2025-09-07T07:36:03.7789396Z * [new branch] gh/bdhirsh/668/base -> origin/gh/bdhirsh/668/base 2025-09-07T07:36:03.7790635Z * [new branch] gh/bdhirsh/668/head -> origin/gh/bdhirsh/668/head 2025-09-07T07:36:03.7791812Z * [new branch] gh/bdhirsh/668/orig -> origin/gh/bdhirsh/668/orig 2025-09-07T07:36:03.7793737Z * [new branch] gh/bdhirsh/669/base -> origin/gh/bdhirsh/669/base 2025-09-07T07:36:03.7808415Z * [new branch] gh/bdhirsh/669/head -> origin/gh/bdhirsh/669/head 2025-09-07T07:36:03.7808785Z * [new branch] gh/bdhirsh/669/orig -> origin/gh/bdhirsh/669/orig 2025-09-07T07:36:03.7808943Z * [new branch] gh/bdhirsh/670/base -> origin/gh/bdhirsh/670/base 2025-09-07T07:36:03.7809085Z * [new branch] gh/bdhirsh/670/head -> origin/gh/bdhirsh/670/head 2025-09-07T07:36:03.7809216Z * [new branch] gh/bdhirsh/670/orig -> origin/gh/bdhirsh/670/orig 2025-09-07T07:36:03.7809387Z * [new branch] gh/benjaminglass1/100/base -> origin/gh/benjaminglass1/100/base 2025-09-07T07:36:03.7809541Z * [new branch] gh/benjaminglass1/100/head -> origin/gh/benjaminglass1/100/head 2025-09-07T07:36:03.7809698Z * [new branch] gh/benjaminglass1/100/orig -> origin/gh/benjaminglass1/100/orig 2025-09-07T07:36:03.7809854Z * [new branch] gh/benjaminglass1/101/base -> origin/gh/benjaminglass1/101/base 2025-09-07T07:36:03.7810007Z * [new branch] gh/benjaminglass1/101/head -> origin/gh/benjaminglass1/101/head 2025-09-07T07:36:03.7810178Z * [new branch] gh/benjaminglass1/101/orig -> origin/gh/benjaminglass1/101/orig 2025-09-07T07:36:03.7811479Z * [new branch] gh/benjaminglass1/102/base -> origin/gh/benjaminglass1/102/base 2025-09-07T07:36:03.7812654Z * [new branch] gh/benjaminglass1/102/head -> origin/gh/benjaminglass1/102/head 2025-09-07T07:36:03.7813844Z * [new branch] gh/benjaminglass1/102/orig -> origin/gh/benjaminglass1/102/orig 2025-09-07T07:36:03.7815601Z * [new branch] gh/benjaminglass1/103/base -> origin/gh/benjaminglass1/103/base 2025-09-07T07:36:03.7816772Z * [new branch] gh/benjaminglass1/103/head -> origin/gh/benjaminglass1/103/head 2025-09-07T07:36:03.7818198Z * [new branch] gh/benjaminglass1/103/orig -> origin/gh/benjaminglass1/103/orig 2025-09-07T07:36:03.7820087Z * [new branch] gh/benjaminglass1/104/base -> origin/gh/benjaminglass1/104/base 2025-09-07T07:36:03.7821150Z * [new branch] gh/benjaminglass1/104/head -> origin/gh/benjaminglass1/104/head 2025-09-07T07:36:03.7822347Z * [new branch] gh/benjaminglass1/104/orig -> origin/gh/benjaminglass1/104/orig 2025-09-07T07:36:03.7824034Z * [new branch] gh/benjaminglass1/105/base -> origin/gh/benjaminglass1/105/base 2025-09-07T07:36:03.7825220Z * [new branch] gh/benjaminglass1/105/head -> origin/gh/benjaminglass1/105/head 2025-09-07T07:36:03.7826419Z * [new branch] gh/benjaminglass1/105/orig -> origin/gh/benjaminglass1/105/orig 2025-09-07T07:36:03.7828249Z * [new branch] gh/benjaminglass1/106/base -> origin/gh/benjaminglass1/106/base 2025-09-07T07:36:03.7829445Z * [new branch] gh/benjaminglass1/106/head -> origin/gh/benjaminglass1/106/head 2025-09-07T07:36:03.7830689Z * [new branch] gh/benjaminglass1/106/orig -> origin/gh/benjaminglass1/106/orig 2025-09-07T07:36:03.7832360Z * [new branch] gh/benjaminglass1/79/base -> origin/gh/benjaminglass1/79/base 2025-09-07T07:36:03.7833535Z * [new branch] gh/benjaminglass1/79/head -> origin/gh/benjaminglass1/79/head 2025-09-07T07:36:03.7834689Z * [new branch] gh/benjaminglass1/79/orig -> origin/gh/benjaminglass1/79/orig 2025-09-07T07:36:03.7836384Z * [new branch] gh/benjaminglass1/86/base -> origin/gh/benjaminglass1/86/base 2025-09-07T07:36:03.7837571Z * [new branch] gh/benjaminglass1/86/head -> origin/gh/benjaminglass1/86/head 2025-09-07T07:36:03.7838863Z * [new branch] gh/benjaminglass1/86/orig -> origin/gh/benjaminglass1/86/orig 2025-09-07T07:36:03.7840567Z * [new branch] gh/benjaminglass1/89/base -> origin/gh/benjaminglass1/89/base 2025-09-07T07:36:03.7841768Z * [new branch] gh/benjaminglass1/89/head -> origin/gh/benjaminglass1/89/head 2025-09-07T07:36:03.7842923Z * [new branch] gh/benjaminglass1/89/orig -> origin/gh/benjaminglass1/89/orig 2025-09-07T07:36:03.7844673Z * [new branch] gh/benjaminglass1/91/base -> origin/gh/benjaminglass1/91/base 2025-09-07T07:36:03.7845891Z * [new branch] gh/benjaminglass1/91/head -> origin/gh/benjaminglass1/91/head 2025-09-07T07:36:03.7847081Z * [new branch] gh/benjaminglass1/91/orig -> origin/gh/benjaminglass1/91/orig 2025-09-07T07:36:03.7848820Z * [new branch] gh/benjaminglass1/93/base -> origin/gh/benjaminglass1/93/base 2025-09-07T07:36:03.7850014Z * [new branch] gh/benjaminglass1/93/head -> origin/gh/benjaminglass1/93/head 2025-09-07T07:36:03.7851269Z * [new branch] gh/benjaminglass1/93/orig -> origin/gh/benjaminglass1/93/orig 2025-09-07T07:36:03.7852937Z * [new branch] gh/benjaminglass1/95/base -> origin/gh/benjaminglass1/95/base 2025-09-07T07:36:03.7854320Z * [new branch] gh/benjaminglass1/95/head -> origin/gh/benjaminglass1/95/head 2025-09-07T07:36:03.7855513Z * [new branch] gh/benjaminglass1/95/orig -> origin/gh/benjaminglass1/95/orig 2025-09-07T07:36:03.7857292Z * [new branch] gh/benjaminglass1/97/base -> origin/gh/benjaminglass1/97/base 2025-09-07T07:36:03.7858564Z * [new branch] gh/benjaminglass1/97/head -> origin/gh/benjaminglass1/97/head 2025-09-07T07:36:03.7859779Z * [new branch] gh/benjaminglass1/97/orig -> origin/gh/benjaminglass1/97/orig 2025-09-07T07:36:03.7861673Z * [new branch] gh/benjaminglass1/99/base -> origin/gh/benjaminglass1/99/base 2025-09-07T07:36:03.7862850Z * [new branch] gh/benjaminglass1/99/head -> origin/gh/benjaminglass1/99/head 2025-09-07T07:36:03.7864179Z * [new branch] gh/benjaminglass1/99/orig -> origin/gh/benjaminglass1/99/orig 2025-09-07T07:36:03.7867352Z * [new branch] gh/bobrenjc93/514/base -> origin/gh/bobrenjc93/514/base 2025-09-07T07:36:03.7867918Z * [new branch] gh/bobrenjc93/514/head -> origin/gh/bobrenjc93/514/head 2025-09-07T07:36:03.7868792Z * [new branch] gh/bobrenjc93/514/orig -> origin/gh/bobrenjc93/514/orig 2025-09-07T07:36:03.7870439Z * [new branch] gh/bobrenjc93/521/base -> origin/gh/bobrenjc93/521/base 2025-09-07T07:36:03.7871624Z * [new branch] gh/bobrenjc93/521/head -> origin/gh/bobrenjc93/521/head 2025-09-07T07:36:03.7872841Z * [new branch] gh/bobrenjc93/521/orig -> origin/gh/bobrenjc93/521/orig 2025-09-07T07:36:03.7874525Z * [new branch] gh/bobrenjc93/522/base -> origin/gh/bobrenjc93/522/base 2025-09-07T07:36:03.7875774Z * [new branch] gh/bobrenjc93/522/head -> origin/gh/bobrenjc93/522/head 2025-09-07T07:36:03.7877025Z * [new branch] gh/bobrenjc93/522/orig -> origin/gh/bobrenjc93/522/orig 2025-09-07T07:36:03.7878712Z * [new branch] gh/bobrenjc93/525/base -> origin/gh/bobrenjc93/525/base 2025-09-07T07:36:03.7879890Z * [new branch] gh/bobrenjc93/525/head -> origin/gh/bobrenjc93/525/head 2025-09-07T07:36:03.7881085Z * [new branch] gh/bobrenjc93/525/orig -> origin/gh/bobrenjc93/525/orig 2025-09-07T07:36:03.7882812Z * [new branch] gh/bobrenjc93/526/base -> origin/gh/bobrenjc93/526/base 2025-09-07T07:36:03.7883991Z * [new branch] gh/bobrenjc93/526/head -> origin/gh/bobrenjc93/526/head 2025-09-07T07:36:03.7885179Z * [new branch] gh/bobrenjc93/526/orig -> origin/gh/bobrenjc93/526/orig 2025-09-07T07:36:03.7886890Z * [new branch] gh/bobrenjc93/527/base -> origin/gh/bobrenjc93/527/base 2025-09-07T07:36:03.7888107Z * [new branch] gh/bobrenjc93/527/head -> origin/gh/bobrenjc93/527/head 2025-09-07T07:36:03.7889451Z * [new branch] gh/bobrenjc93/527/orig -> origin/gh/bobrenjc93/527/orig 2025-09-07T07:36:03.7891153Z * [new branch] gh/bobrenjc93/528/base -> origin/gh/bobrenjc93/528/base 2025-09-07T07:36:03.7892376Z * [new branch] gh/bobrenjc93/528/head -> origin/gh/bobrenjc93/528/head 2025-09-07T07:36:03.7893610Z * [new branch] gh/bobrenjc93/528/orig -> origin/gh/bobrenjc93/528/orig 2025-09-07T07:36:03.7895349Z * [new branch] gh/bobrenjc93/529/base -> origin/gh/bobrenjc93/529/base 2025-09-07T07:36:03.7896507Z * [new branch] gh/bobrenjc93/529/head -> origin/gh/bobrenjc93/529/head 2025-09-07T07:36:03.7897872Z * [new branch] gh/bobrenjc93/529/orig -> origin/gh/bobrenjc93/529/orig 2025-09-07T07:36:03.7899614Z * [new branch] gh/bobrenjc93/535/base -> origin/gh/bobrenjc93/535/base 2025-09-07T07:36:03.7900849Z * [new branch] gh/bobrenjc93/535/head -> origin/gh/bobrenjc93/535/head 2025-09-07T07:36:03.7902020Z * [new branch] gh/bobrenjc93/535/orig -> origin/gh/bobrenjc93/535/orig 2025-09-07T07:36:03.7904106Z * [new branch] gh/bobrenjc93/537/base -> origin/gh/bobrenjc93/537/base 2025-09-07T07:36:03.7905439Z * [new branch] gh/bobrenjc93/537/head -> origin/gh/bobrenjc93/537/head 2025-09-07T07:36:03.7906824Z * [new branch] gh/bobrenjc93/537/orig -> origin/gh/bobrenjc93/537/orig 2025-09-07T07:36:03.7908766Z * [new branch] gh/bobrenjc93/539/base -> origin/gh/bobrenjc93/539/base 2025-09-07T07:36:03.7910002Z * [new branch] gh/bobrenjc93/539/head -> origin/gh/bobrenjc93/539/head 2025-09-07T07:36:03.7911319Z * [new branch] gh/bobrenjc93/539/orig -> origin/gh/bobrenjc93/539/orig 2025-09-07T07:36:03.7913087Z * [new branch] gh/bobrenjc93/540/base -> origin/gh/bobrenjc93/540/base 2025-09-07T07:36:03.7914344Z * [new branch] gh/bobrenjc93/540/head -> origin/gh/bobrenjc93/540/head 2025-09-07T07:36:03.7915713Z * [new branch] gh/bobrenjc93/540/orig -> origin/gh/bobrenjc93/540/orig 2025-09-07T07:36:03.7917344Z * [new branch] gh/bobrenjc93/541/base -> origin/gh/bobrenjc93/541/base 2025-09-07T07:36:03.7918548Z * [new branch] gh/bobrenjc93/541/head -> origin/gh/bobrenjc93/541/head 2025-09-07T07:36:03.7919804Z * [new branch] gh/bobrenjc93/541/orig -> origin/gh/bobrenjc93/541/orig 2025-09-07T07:36:03.7921408Z * [new branch] gh/bobrenjc93/542/base -> origin/gh/bobrenjc93/542/base 2025-09-07T07:36:03.7922571Z * [new branch] gh/bobrenjc93/542/head -> origin/gh/bobrenjc93/542/head 2025-09-07T07:36:03.7923780Z * [new branch] gh/bobrenjc93/542/orig -> origin/gh/bobrenjc93/542/orig 2025-09-07T07:36:03.7925597Z * [new branch] gh/bobrenjc93/543/base -> origin/gh/bobrenjc93/543/base 2025-09-07T07:36:03.7926789Z * [new branch] gh/bobrenjc93/543/head -> origin/gh/bobrenjc93/543/head 2025-09-07T07:36:03.7928057Z * [new branch] gh/bobrenjc93/543/orig -> origin/gh/bobrenjc93/543/orig 2025-09-07T07:36:03.7929684Z * [new branch] gh/bobrenjc93/544/base -> origin/gh/bobrenjc93/544/base 2025-09-07T07:36:03.7930881Z * [new branch] gh/bobrenjc93/544/head -> origin/gh/bobrenjc93/544/head 2025-09-07T07:36:03.7932104Z * [new branch] gh/bobrenjc93/544/orig -> origin/gh/bobrenjc93/544/orig 2025-09-07T07:36:03.7934192Z * [new branch] gh/bobrenjc93/545/base -> origin/gh/bobrenjc93/545/base 2025-09-07T07:36:03.7935584Z * [new branch] gh/bobrenjc93/545/head -> origin/gh/bobrenjc93/545/head 2025-09-07T07:36:03.7936768Z * [new branch] gh/bobrenjc93/545/orig -> origin/gh/bobrenjc93/545/orig 2025-09-07T07:36:03.7938641Z * [new branch] gh/bobrenjc93/546/base -> origin/gh/bobrenjc93/546/base 2025-09-07T07:36:03.7939841Z * [new branch] gh/bobrenjc93/546/head -> origin/gh/bobrenjc93/546/head 2025-09-07T07:36:03.7941029Z * [new branch] gh/bobrenjc93/546/orig -> origin/gh/bobrenjc93/546/orig 2025-09-07T07:36:03.7943315Z * [new branch] gh/bobrenjc93/547/base -> origin/gh/bobrenjc93/547/base 2025-09-07T07:36:03.7944563Z * [new branch] gh/bobrenjc93/547/head -> origin/gh/bobrenjc93/547/head 2025-09-07T07:36:03.7945867Z * [new branch] gh/bobrenjc93/547/orig -> origin/gh/bobrenjc93/547/orig 2025-09-07T07:36:03.7947516Z * [new branch] gh/bobrenjc93/548/base -> origin/gh/bobrenjc93/548/base 2025-09-07T07:36:03.7948695Z * [new branch] gh/bobrenjc93/548/head -> origin/gh/bobrenjc93/548/head 2025-09-07T07:36:03.7949865Z * [new branch] gh/bobrenjc93/548/orig -> origin/gh/bobrenjc93/548/orig 2025-09-07T07:36:03.7951512Z * [new branch] gh/bobrenjc93/549/base -> origin/gh/bobrenjc93/549/base 2025-09-07T07:36:03.7952759Z * [new branch] gh/bobrenjc93/549/head -> origin/gh/bobrenjc93/549/head 2025-09-07T07:36:03.7954022Z * [new branch] gh/bobrenjc93/549/orig -> origin/gh/bobrenjc93/549/orig 2025-09-07T07:36:03.7955972Z * [new branch] gh/bobrenjc93/550/base -> origin/gh/bobrenjc93/550/base 2025-09-07T07:36:03.7957144Z * [new branch] gh/bobrenjc93/550/head -> origin/gh/bobrenjc93/550/head 2025-09-07T07:36:03.7958357Z * [new branch] gh/bobrenjc93/550/orig -> origin/gh/bobrenjc93/550/orig 2025-09-07T07:36:03.7960290Z * [new branch] gh/bobrenjc93/551/base -> origin/gh/bobrenjc93/551/base 2025-09-07T07:36:03.7961634Z * [new branch] gh/bobrenjc93/551/head -> origin/gh/bobrenjc93/551/head 2025-09-07T07:36:03.7962899Z * [new branch] gh/bobrenjc93/551/orig -> origin/gh/bobrenjc93/551/orig 2025-09-07T07:36:03.7964896Z * [new branch] gh/bobrenjc93/552/base -> origin/gh/bobrenjc93/552/base 2025-09-07T07:36:03.7965957Z * [new branch] gh/bobrenjc93/552/head -> origin/gh/bobrenjc93/552/head 2025-09-07T07:36:03.7967375Z * [new branch] gh/bobrenjc93/552/orig -> origin/gh/bobrenjc93/552/orig 2025-09-07T07:36:03.7968929Z * [new branch] gh/bobrenjc93/553/base -> origin/gh/bobrenjc93/553/base 2025-09-07T07:36:03.7970185Z * [new branch] gh/bobrenjc93/553/head -> origin/gh/bobrenjc93/553/head 2025-09-07T07:36:03.7971349Z * [new branch] gh/bobrenjc93/553/orig -> origin/gh/bobrenjc93/553/orig 2025-09-07T07:36:03.7972991Z * [new branch] gh/bobrenjc93/554/base -> origin/gh/bobrenjc93/554/base 2025-09-07T07:36:03.7974185Z * [new branch] gh/bobrenjc93/554/head -> origin/gh/bobrenjc93/554/head 2025-09-07T07:36:03.7975367Z * [new branch] gh/bobrenjc93/554/orig -> origin/gh/bobrenjc93/554/orig 2025-09-07T07:36:03.7977177Z * [new branch] gh/bobrenjc93/555/base -> origin/gh/bobrenjc93/555/base 2025-09-07T07:36:03.7978406Z * [new branch] gh/bobrenjc93/555/head -> origin/gh/bobrenjc93/555/head 2025-09-07T07:36:03.7979688Z * [new branch] gh/bobrenjc93/555/orig -> origin/gh/bobrenjc93/555/orig 2025-09-07T07:36:03.7981443Z * [new branch] gh/bobrenjc93/556/base -> origin/gh/bobrenjc93/556/base 2025-09-07T07:36:03.7982657Z * [new branch] gh/bobrenjc93/556/head -> origin/gh/bobrenjc93/556/head 2025-09-07T07:36:03.7983875Z * [new branch] gh/bobrenjc93/556/orig -> origin/gh/bobrenjc93/556/orig 2025-09-07T07:36:03.7986139Z * [new branch] gh/briancoutinho/2/base -> origin/gh/briancoutinho/2/base 2025-09-07T07:36:03.7987482Z * [new branch] gh/briancoutinho/2/head -> origin/gh/briancoutinho/2/head 2025-09-07T07:36:03.7989670Z * [new branch] gh/c00w/23/base -> origin/gh/c00w/23/base 2025-09-07T07:36:03.7990853Z * [new branch] gh/c00w/23/head -> origin/gh/c00w/23/head 2025-09-07T07:36:03.7992929Z * [new branch] gh/c00w/48/base -> origin/gh/c00w/48/base 2025-09-07T07:36:03.7993820Z * [new branch] gh/c00w/48/head -> origin/gh/c00w/48/head 2025-09-07T07:36:03.7995095Z * [new branch] gh/c00w/48/orig -> origin/gh/c00w/48/orig 2025-09-07T07:36:03.7996935Z * [new branch] gh/c00w/53/base -> origin/gh/c00w/53/base 2025-09-07T07:36:03.7998254Z * [new branch] gh/c00w/53/head -> origin/gh/c00w/53/head 2025-09-07T07:36:03.7999459Z * [new branch] gh/c00w/53/orig -> origin/gh/c00w/53/orig 2025-09-07T07:36:03.8001055Z * [new branch] gh/c00w/54/base -> origin/gh/c00w/54/base 2025-09-07T07:36:03.8002293Z * [new branch] gh/c00w/54/head -> origin/gh/c00w/54/head 2025-09-07T07:36:03.8003711Z * [new branch] gh/c00w/54/orig -> origin/gh/c00w/54/orig 2025-09-07T07:36:03.8005545Z * [new branch] gh/c00w/55/base -> origin/gh/c00w/55/base 2025-09-07T07:36:03.8006812Z * [new branch] gh/c00w/55/head -> origin/gh/c00w/55/head 2025-09-07T07:36:03.8008188Z * [new branch] gh/c00w/55/orig -> origin/gh/c00w/55/orig 2025-09-07T07:36:03.8009930Z * [new branch] gh/c00w/56/base -> origin/gh/c00w/56/base 2025-09-07T07:36:03.8011357Z * [new branch] gh/c00w/56/head -> origin/gh/c00w/56/head 2025-09-07T07:36:03.8012749Z * [new branch] gh/c00w/56/orig -> origin/gh/c00w/56/orig 2025-09-07T07:36:03.8014879Z * [new branch] gh/clee2000/1/base -> origin/gh/clee2000/1/base 2025-09-07T07:36:03.8016320Z * [new branch] gh/clee2000/1/head -> origin/gh/clee2000/1/head 2025-09-07T07:36:03.8017374Z * [new branch] gh/clee2000/1/orig -> origin/gh/clee2000/1/orig 2025-09-07T07:36:03.8019695Z * [new branch] gh/coconutruben/1/base -> origin/gh/coconutruben/1/base 2025-09-07T07:36:03.8020976Z * [new branch] gh/coconutruben/1/head -> origin/gh/coconutruben/1/head 2025-09-07T07:36:03.8022821Z * [new branch] gh/coconutruben/11/base -> origin/gh/coconutruben/11/base 2025-09-07T07:36:03.8024150Z * [new branch] gh/coconutruben/11/head -> origin/gh/coconutruben/11/head 2025-09-07T07:36:03.8025410Z * [new branch] gh/coconutruben/11/orig -> origin/gh/coconutruben/11/orig 2025-09-07T07:36:03.8027508Z * [new branch] gh/coconutruben/12/base -> origin/gh/coconutruben/12/base 2025-09-07T07:36:03.8028959Z * [new branch] gh/coconutruben/12/head -> origin/gh/coconutruben/12/head 2025-09-07T07:36:03.8030383Z * [new branch] gh/coconutruben/12/orig -> origin/gh/coconutruben/12/orig 2025-09-07T07:36:03.8032189Z * [new branch] gh/coconutruben/13/base -> origin/gh/coconutruben/13/base 2025-09-07T07:36:03.8033480Z * [new branch] gh/coconutruben/13/head -> origin/gh/coconutruben/13/head 2025-09-07T07:36:03.8034824Z * [new branch] gh/coconutruben/13/orig -> origin/gh/coconutruben/13/orig 2025-09-07T07:36:03.8036740Z * [new branch] gh/coconutruben/14/base -> origin/gh/coconutruben/14/base 2025-09-07T07:36:03.8038005Z * [new branch] gh/coconutruben/14/head -> origin/gh/coconutruben/14/head 2025-09-07T07:36:03.8039201Z * [new branch] gh/coconutruben/14/orig -> origin/gh/coconutruben/14/orig 2025-09-07T07:36:03.8041153Z * [new branch] gh/coconutruben/15/base -> origin/gh/coconutruben/15/base 2025-09-07T07:36:03.8042498Z * [new branch] gh/coconutruben/15/head -> origin/gh/coconutruben/15/head 2025-09-07T07:36:03.8043788Z * [new branch] gh/coconutruben/15/orig -> origin/gh/coconutruben/15/orig 2025-09-07T07:36:03.8045454Z * [new branch] gh/coconutruben/16/base -> origin/gh/coconutruben/16/base 2025-09-07T07:36:03.8046663Z * [new branch] gh/coconutruben/16/head -> origin/gh/coconutruben/16/head 2025-09-07T07:36:03.8047842Z * [new branch] gh/coconutruben/16/orig -> origin/gh/coconutruben/16/orig 2025-09-07T07:36:03.8049882Z * [new branch] gh/coconutruben/17/base -> origin/gh/coconutruben/17/base 2025-09-07T07:36:03.8051195Z * [new branch] gh/coconutruben/17/head -> origin/gh/coconutruben/17/head 2025-09-07T07:36:03.8052544Z * [new branch] gh/coconutruben/17/orig -> origin/gh/coconutruben/17/orig 2025-09-07T07:36:03.8054313Z * [new branch] gh/coconutruben/18/base -> origin/gh/coconutruben/18/base 2025-09-07T07:36:03.8055657Z * [new branch] gh/coconutruben/18/head -> origin/gh/coconutruben/18/head 2025-09-07T07:36:03.8056826Z * [new branch] gh/coconutruben/18/orig -> origin/gh/coconutruben/18/orig 2025-09-07T07:36:03.8058815Z * [new branch] gh/coconutruben/19/base -> origin/gh/coconutruben/19/base 2025-09-07T07:36:03.8060127Z * [new branch] gh/coconutruben/19/head -> origin/gh/coconutruben/19/head 2025-09-07T07:36:03.8061381Z * [new branch] gh/coconutruben/19/orig -> origin/gh/coconutruben/19/orig 2025-09-07T07:36:03.8063270Z * [new branch] gh/coconutruben/20/base -> origin/gh/coconutruben/20/base 2025-09-07T07:36:03.8064513Z * [new branch] gh/coconutruben/20/head -> origin/gh/coconutruben/20/head 2025-09-07T07:36:03.8065823Z * [new branch] gh/coconutruben/20/orig -> origin/gh/coconutruben/20/orig 2025-09-07T07:36:03.8067703Z * [new branch] gh/coconutruben/21/base -> origin/gh/coconutruben/21/base 2025-09-07T07:36:03.8068969Z * [new branch] gh/coconutruben/21/head -> origin/gh/coconutruben/21/head 2025-09-07T07:36:03.8070089Z * [new branch] gh/coconutruben/21/orig -> origin/gh/coconutruben/21/orig 2025-09-07T07:36:03.8071860Z * [new branch] gh/coconutruben/22/base -> origin/gh/coconutruben/22/base 2025-09-07T07:36:03.8073161Z * [new branch] gh/coconutruben/22/head -> origin/gh/coconutruben/22/head 2025-09-07T07:36:03.8074512Z * [new branch] gh/coconutruben/22/orig -> origin/gh/coconutruben/22/orig 2025-09-07T07:36:03.8076312Z * [new branch] gh/coconutruben/24/base -> origin/gh/coconutruben/24/base 2025-09-07T07:36:03.8077595Z * [new branch] gh/coconutruben/24/head -> origin/gh/coconutruben/24/head 2025-09-07T07:36:03.8078839Z * [new branch] gh/coconutruben/24/orig -> origin/gh/coconutruben/24/orig 2025-09-07T07:36:03.8080954Z * [new branch] gh/coconutruben/25/base -> origin/gh/coconutruben/25/base 2025-09-07T07:36:03.8082432Z * [new branch] gh/coconutruben/25/head -> origin/gh/coconutruben/25/head 2025-09-07T07:36:03.8083837Z * [new branch] gh/coconutruben/25/orig -> origin/gh/coconutruben/25/orig 2025-09-07T07:36:03.8085721Z * [new branch] gh/coconutruben/28/base -> origin/gh/coconutruben/28/base 2025-09-07T07:36:03.8086968Z * [new branch] gh/coconutruben/28/head -> origin/gh/coconutruben/28/head 2025-09-07T07:36:03.8088235Z * [new branch] gh/coconutruben/28/orig -> origin/gh/coconutruben/28/orig 2025-09-07T07:36:03.8090086Z * [new branch] gh/coconutruben/29/base -> origin/gh/coconutruben/29/base 2025-09-07T07:36:03.8091338Z * [new branch] gh/coconutruben/29/head -> origin/gh/coconutruben/29/head 2025-09-07T07:36:03.8092562Z * [new branch] gh/coconutruben/29/orig -> origin/gh/coconutruben/29/orig 2025-09-07T07:36:03.8094392Z * [new branch] gh/coconutruben/30/base -> origin/gh/coconutruben/30/base 2025-09-07T07:36:03.8095647Z * [new branch] gh/coconutruben/30/head -> origin/gh/coconutruben/30/head 2025-09-07T07:36:03.8096964Z * [new branch] gh/coconutruben/30/orig -> origin/gh/coconutruben/30/orig 2025-09-07T07:36:03.8099304Z * [new branch] gh/coconutruben/31/base -> origin/gh/coconutruben/31/base 2025-09-07T07:36:03.8100647Z * [new branch] gh/coconutruben/31/head -> origin/gh/coconutruben/31/head 2025-09-07T07:36:03.8101926Z * [new branch] gh/coconutruben/31/orig -> origin/gh/coconutruben/31/orig 2025-09-07T07:36:03.8108113Z * [new branch] gh/coconutruben/32/base -> origin/gh/coconutruben/32/base 2025-09-07T07:36:03.8109471Z * [new branch] gh/coconutruben/32/head -> origin/gh/coconutruben/32/head 2025-09-07T07:36:03.8110704Z * [new branch] gh/coconutruben/32/orig -> origin/gh/coconutruben/32/orig 2025-09-07T07:36:03.8112712Z * [new branch] gh/coconutruben/33/base -> origin/gh/coconutruben/33/base 2025-09-07T07:36:03.8114113Z * [new branch] gh/coconutruben/33/head -> origin/gh/coconutruben/33/head 2025-09-07T07:36:03.8115465Z * [new branch] gh/coconutruben/33/orig -> origin/gh/coconutruben/33/orig 2025-09-07T07:36:03.8117553Z * [new branch] gh/coconutruben/34/base -> origin/gh/coconutruben/34/base 2025-09-07T07:36:03.8118768Z * [new branch] gh/coconutruben/34/head -> origin/gh/coconutruben/34/head 2025-09-07T07:36:03.8119914Z * [new branch] gh/coconutruben/34/orig -> origin/gh/coconutruben/34/orig 2025-09-07T07:36:03.8121717Z * [new branch] gh/coconutruben/35/base -> origin/gh/coconutruben/35/base 2025-09-07T07:36:03.8122983Z * [new branch] gh/coconutruben/35/head -> origin/gh/coconutruben/35/head 2025-09-07T07:36:03.8124409Z * [new branch] gh/coconutruben/35/orig -> origin/gh/coconutruben/35/orig 2025-09-07T07:36:03.8127223Z * [new branch] gh/coconutruben/36/base -> origin/gh/coconutruben/36/base 2025-09-07T07:36:03.8128832Z * [new branch] gh/coconutruben/36/head -> origin/gh/coconutruben/36/head 2025-09-07T07:36:03.8130620Z * [new branch] gh/coconutruben/36/orig -> origin/gh/coconutruben/36/orig 2025-09-07T07:36:03.8132622Z * [new branch] gh/coconutruben/37/base -> origin/gh/coconutruben/37/base 2025-09-07T07:36:03.8133871Z * [new branch] gh/coconutruben/37/head -> origin/gh/coconutruben/37/head 2025-09-07T07:36:03.8135113Z * [new branch] gh/coconutruben/37/orig -> origin/gh/coconutruben/37/orig 2025-09-07T07:36:03.8137006Z * [new branch] gh/coconutruben/38/base -> origin/gh/coconutruben/38/base 2025-09-07T07:36:03.8138458Z * [new branch] gh/coconutruben/38/head -> origin/gh/coconutruben/38/head 2025-09-07T07:36:03.8139694Z * [new branch] gh/coconutruben/38/orig -> origin/gh/coconutruben/38/orig 2025-09-07T07:36:03.8141624Z * [new branch] gh/coconutruben/39/base -> origin/gh/coconutruben/39/base 2025-09-07T07:36:03.8142780Z * [new branch] gh/coconutruben/39/head -> origin/gh/coconutruben/39/head 2025-09-07T07:36:03.8144001Z * [new branch] gh/coconutruben/39/orig -> origin/gh/coconutruben/39/orig 2025-09-07T07:36:03.8146077Z * [new branch] gh/coconutruben/40/base -> origin/gh/coconutruben/40/base 2025-09-07T07:36:03.8147235Z * [new branch] gh/coconutruben/40/head -> origin/gh/coconutruben/40/head 2025-09-07T07:36:03.8148456Z * [new branch] gh/coconutruben/40/orig -> origin/gh/coconutruben/40/orig 2025-09-07T07:36:03.8150523Z * [new branch] gh/coconutruben/41/base -> origin/gh/coconutruben/41/base 2025-09-07T07:36:03.8151776Z * [new branch] gh/coconutruben/41/head -> origin/gh/coconutruben/41/head 2025-09-07T07:36:03.8153031Z * [new branch] gh/coconutruben/41/orig -> origin/gh/coconutruben/41/orig 2025-09-07T07:36:03.8154965Z * [new branch] gh/coconutruben/42/base -> origin/gh/coconutruben/42/base 2025-09-07T07:36:03.8156342Z * [new branch] gh/coconutruben/42/head -> origin/gh/coconutruben/42/head 2025-09-07T07:36:03.8157597Z * [new branch] gh/coconutruben/42/orig -> origin/gh/coconutruben/42/orig 2025-09-07T07:36:03.8159577Z * [new branch] gh/coconutruben/43/base -> origin/gh/coconutruben/43/base 2025-09-07T07:36:03.8160834Z * [new branch] gh/coconutruben/43/head -> origin/gh/coconutruben/43/head 2025-09-07T07:36:03.8162055Z * [new branch] gh/coconutruben/43/orig -> origin/gh/coconutruben/43/orig 2025-09-07T07:36:03.8164103Z * [new branch] gh/coconutruben/44/base -> origin/gh/coconutruben/44/base 2025-09-07T07:36:03.8165385Z * [new branch] gh/coconutruben/44/head -> origin/gh/coconutruben/44/head 2025-09-07T07:36:03.8166668Z * [new branch] gh/coconutruben/44/orig -> origin/gh/coconutruben/44/orig 2025-09-07T07:36:03.8168575Z * [new branch] gh/coconutruben/45/base -> origin/gh/coconutruben/45/base 2025-09-07T07:36:03.8169830Z * [new branch] gh/coconutruben/45/head -> origin/gh/coconutruben/45/head 2025-09-07T07:36:03.8171119Z * [new branch] gh/coconutruben/45/orig -> origin/gh/coconutruben/45/orig 2025-09-07T07:36:03.8172913Z * [new branch] gh/coconutruben/46/base -> origin/gh/coconutruben/46/base 2025-09-07T07:36:03.8174201Z * [new branch] gh/coconutruben/46/head -> origin/gh/coconutruben/46/head 2025-09-07T07:36:03.8175599Z * [new branch] gh/coconutruben/46/orig -> origin/gh/coconutruben/46/orig 2025-09-07T07:36:03.8177463Z * [new branch] gh/coconutruben/47/base -> origin/gh/coconutruben/47/base 2025-09-07T07:36:03.8178763Z * [new branch] gh/coconutruben/47/head -> origin/gh/coconutruben/47/head 2025-09-07T07:36:03.8180027Z * [new branch] gh/coconutruben/47/orig -> origin/gh/coconutruben/47/orig 2025-09-07T07:36:03.8182153Z * [new branch] gh/coconutruben/48/base -> origin/gh/coconutruben/48/base 2025-09-07T07:36:03.8183480Z * [new branch] gh/coconutruben/48/head -> origin/gh/coconutruben/48/head 2025-09-07T07:36:03.8184712Z * [new branch] gh/coconutruben/48/orig -> origin/gh/coconutruben/48/orig 2025-09-07T07:36:03.8186742Z * [new branch] gh/coconutruben/49/base -> origin/gh/coconutruben/49/base 2025-09-07T07:36:03.8187997Z * [new branch] gh/coconutruben/49/head -> origin/gh/coconutruben/49/head 2025-09-07T07:36:03.8189220Z * [new branch] gh/coconutruben/49/orig -> origin/gh/coconutruben/49/orig 2025-09-07T07:36:03.8191078Z * [new branch] gh/coconutruben/50/base -> origin/gh/coconutruben/50/base 2025-09-07T07:36:03.8192400Z * [new branch] gh/coconutruben/50/head -> origin/gh/coconutruben/50/head 2025-09-07T07:36:03.8193773Z * [new branch] gh/coconutruben/50/orig -> origin/gh/coconutruben/50/orig 2025-09-07T07:36:03.8195548Z * [new branch] gh/coconutruben/51/base -> origin/gh/coconutruben/51/base 2025-09-07T07:36:03.8196844Z * [new branch] gh/coconutruben/51/head -> origin/gh/coconutruben/51/head 2025-09-07T07:36:03.8198084Z * [new branch] gh/coconutruben/51/orig -> origin/gh/coconutruben/51/orig 2025-09-07T07:36:03.8200130Z * [new branch] gh/coconutruben/52/base -> origin/gh/coconutruben/52/base 2025-09-07T07:36:03.8201409Z * [new branch] gh/coconutruben/52/head -> origin/gh/coconutruben/52/head 2025-09-07T07:36:03.8202705Z * [new branch] gh/coconutruben/52/orig -> origin/gh/coconutruben/52/orig 2025-09-07T07:36:03.8204762Z * [new branch] gh/coconutruben/53/base -> origin/gh/coconutruben/53/base 2025-09-07T07:36:03.8205941Z * [new branch] gh/coconutruben/53/head -> origin/gh/coconutruben/53/head 2025-09-07T07:36:03.8207171Z * [new branch] gh/coconutruben/53/orig -> origin/gh/coconutruben/53/orig 2025-09-07T07:36:03.8209016Z * [new branch] gh/coconutruben/54/base -> origin/gh/coconutruben/54/base 2025-09-07T07:36:03.8210295Z * [new branch] gh/coconutruben/54/head -> origin/gh/coconutruben/54/head 2025-09-07T07:36:03.8211519Z * [new branch] gh/coconutruben/54/orig -> origin/gh/coconutruben/54/orig 2025-09-07T07:36:03.8213398Z * [new branch] gh/coconutruben/55/base -> origin/gh/coconutruben/55/base 2025-09-07T07:36:03.8214592Z * [new branch] gh/coconutruben/55/head -> origin/gh/coconutruben/55/head 2025-09-07T07:36:03.8215885Z * [new branch] gh/coconutruben/55/orig -> origin/gh/coconutruben/55/orig 2025-09-07T07:36:03.8218033Z * [new branch] gh/coconutruben/56/base -> origin/gh/coconutruben/56/base 2025-09-07T07:36:03.8219288Z * [new branch] gh/coconutruben/56/head -> origin/gh/coconutruben/56/head 2025-09-07T07:36:03.8220514Z * [new branch] gh/coconutruben/56/orig -> origin/gh/coconutruben/56/orig 2025-09-07T07:36:03.8222398Z * [new branch] gh/coconutruben/57/base -> origin/gh/coconutruben/57/base 2025-09-07T07:36:03.8223693Z * [new branch] gh/coconutruben/57/head -> origin/gh/coconutruben/57/head 2025-09-07T07:36:03.8224960Z * [new branch] gh/coconutruben/57/orig -> origin/gh/coconutruben/57/orig 2025-09-07T07:36:03.8227026Z * [new branch] gh/coconutruben/58/base -> origin/gh/coconutruben/58/base 2025-09-07T07:36:03.8228563Z * [new branch] gh/coconutruben/58/head -> origin/gh/coconutruben/58/head 2025-09-07T07:36:03.8229675Z * [new branch] gh/coconutruben/58/orig -> origin/gh/coconutruben/58/orig 2025-09-07T07:36:03.8231722Z * [new branch] gh/coconutruben/59/base -> origin/gh/coconutruben/59/base 2025-09-07T07:36:03.8232865Z * [new branch] gh/coconutruben/59/head -> origin/gh/coconutruben/59/head 2025-09-07T07:36:03.8234036Z * [new branch] gh/coconutruben/59/orig -> origin/gh/coconutruben/59/orig 2025-09-07T07:36:03.8235921Z * [new branch] gh/coconutruben/60/base -> origin/gh/coconutruben/60/base 2025-09-07T07:36:03.8237173Z * [new branch] gh/coconutruben/60/head -> origin/gh/coconutruben/60/head 2025-09-07T07:36:03.8238474Z * [new branch] gh/coconutruben/60/orig -> origin/gh/coconutruben/60/orig 2025-09-07T07:36:03.8240233Z * [new branch] gh/coconutruben/61/base -> origin/gh/coconutruben/61/base 2025-09-07T07:36:03.8241571Z * [new branch] gh/coconutruben/61/head -> origin/gh/coconutruben/61/head 2025-09-07T07:36:03.8242767Z * [new branch] gh/coconutruben/61/orig -> origin/gh/coconutruben/61/orig 2025-09-07T07:36:03.8244690Z * [new branch] gh/coconutruben/62/base -> origin/gh/coconutruben/62/base 2025-09-07T07:36:03.8245929Z * [new branch] gh/coconutruben/62/head -> origin/gh/coconutruben/62/head 2025-09-07T07:36:03.8247218Z * [new branch] gh/coconutruben/62/orig -> origin/gh/coconutruben/62/orig 2025-09-07T07:36:03.8249166Z * [new branch] gh/coconutruben/63/base -> origin/gh/coconutruben/63/base 2025-09-07T07:36:03.8250424Z * [new branch] gh/coconutruben/63/head -> origin/gh/coconutruben/63/head 2025-09-07T07:36:03.8251655Z * [new branch] gh/coconutruben/63/orig -> origin/gh/coconutruben/63/orig 2025-09-07T07:36:03.8253564Z * [new branch] gh/coconutruben/64/base -> origin/gh/coconutruben/64/base 2025-09-07T07:36:03.8254837Z * [new branch] gh/coconutruben/64/head -> origin/gh/coconutruben/64/head 2025-09-07T07:36:03.8256056Z * [new branch] gh/coconutruben/64/orig -> origin/gh/coconutruben/64/orig 2025-09-07T07:36:03.8257976Z * [new branch] gh/coconutruben/65/base -> origin/gh/coconutruben/65/base 2025-09-07T07:36:03.8259454Z * [new branch] gh/coconutruben/65/head -> origin/gh/coconutruben/65/head 2025-09-07T07:36:03.8260469Z * [new branch] gh/coconutruben/65/orig -> origin/gh/coconutruben/65/orig 2025-09-07T07:36:03.8262283Z * [new branch] gh/coconutruben/66/base -> origin/gh/coconutruben/66/base 2025-09-07T07:36:03.8263491Z * [new branch] gh/coconutruben/66/head -> origin/gh/coconutruben/66/head 2025-09-07T07:36:03.8264668Z * [new branch] gh/coconutruben/66/orig -> origin/gh/coconutruben/66/orig 2025-09-07T07:36:03.8267170Z * [new branch] gh/codingwithsurya/12/base -> origin/gh/codingwithsurya/12/base 2025-09-07T07:36:03.8268688Z * [new branch] gh/codingwithsurya/12/head -> origin/gh/codingwithsurya/12/head 2025-09-07T07:36:03.8270115Z * [new branch] gh/codingwithsurya/12/orig -> origin/gh/codingwithsurya/12/orig 2025-09-07T07:36:03.8271798Z * [new branch] gh/codingwithsurya/14/base -> origin/gh/codingwithsurya/14/base 2025-09-07T07:36:03.8273028Z * [new branch] gh/codingwithsurya/14/head -> origin/gh/codingwithsurya/14/head 2025-09-07T07:36:03.8274301Z * [new branch] gh/codingwithsurya/14/orig -> origin/gh/codingwithsurya/14/orig 2025-09-07T07:36:03.8276201Z * [new branch] gh/codingwithsurya/15/base -> origin/gh/codingwithsurya/15/base 2025-09-07T07:36:03.8277471Z * [new branch] gh/codingwithsurya/15/head -> origin/gh/codingwithsurya/15/head 2025-09-07T07:36:03.8278733Z * [new branch] gh/codingwithsurya/15/orig -> origin/gh/codingwithsurya/15/orig 2025-09-07T07:36:03.8280598Z * [new branch] gh/codingwithsurya/16/base -> origin/gh/codingwithsurya/16/base 2025-09-07T07:36:03.8281810Z * [new branch] gh/codingwithsurya/16/head -> origin/gh/codingwithsurya/16/head 2025-09-07T07:36:03.8283055Z * [new branch] gh/codingwithsurya/16/orig -> origin/gh/codingwithsurya/16/orig 2025-09-07T07:36:03.8284958Z * [new branch] gh/codingwithsurya/17/base -> origin/gh/codingwithsurya/17/base 2025-09-07T07:36:03.8286261Z * [new branch] gh/codingwithsurya/17/head -> origin/gh/codingwithsurya/17/head 2025-09-07T07:36:03.8287524Z * [new branch] gh/codingwithsurya/17/orig -> origin/gh/codingwithsurya/17/orig 2025-09-07T07:36:03.8289330Z * [new branch] gh/codingwithsurya/18/base -> origin/gh/codingwithsurya/18/base 2025-09-07T07:36:03.8290603Z * [new branch] gh/codingwithsurya/18/head -> origin/gh/codingwithsurya/18/head 2025-09-07T07:36:03.8291813Z * [new branch] gh/codingwithsurya/18/orig -> origin/gh/codingwithsurya/18/orig 2025-09-07T07:36:03.8293722Z * [new branch] gh/codingwithsurya/19/base -> origin/gh/codingwithsurya/19/base 2025-09-07T07:36:03.8294924Z * [new branch] gh/codingwithsurya/19/head -> origin/gh/codingwithsurya/19/head 2025-09-07T07:36:03.8296180Z * [new branch] gh/codingwithsurya/19/orig -> origin/gh/codingwithsurya/19/orig 2025-09-07T07:36:03.8298049Z * [new branch] gh/codingwithsurya/20/base -> origin/gh/codingwithsurya/20/base 2025-09-07T07:36:03.8299266Z * [new branch] gh/codingwithsurya/20/head -> origin/gh/codingwithsurya/20/head 2025-09-07T07:36:03.8300460Z * [new branch] gh/codingwithsurya/20/orig -> origin/gh/codingwithsurya/20/orig 2025-09-07T07:36:03.8302319Z * [new branch] gh/codingwithsurya/21/base -> origin/gh/codingwithsurya/21/base 2025-09-07T07:36:03.8303855Z * [new branch] gh/codingwithsurya/21/head -> origin/gh/codingwithsurya/21/head 2025-09-07T07:36:03.8305374Z * [new branch] gh/codingwithsurya/21/orig -> origin/gh/codingwithsurya/21/orig 2025-09-07T07:36:03.8307578Z * [new branch] gh/colinchan15/1/base -> origin/gh/colinchan15/1/base 2025-09-07T07:36:03.8308776Z * [new branch] gh/colinchan15/1/head -> origin/gh/colinchan15/1/head 2025-09-07T07:36:03.8310355Z * [new branch] gh/colinchan15/2/base -> origin/gh/colinchan15/2/base 2025-09-07T07:36:03.8311482Z * [new branch] gh/colinchan15/2/head -> origin/gh/colinchan15/2/head 2025-09-07T07:36:03.8313104Z * [new branch] gh/colinchan15/3/base -> origin/gh/colinchan15/3/base 2025-09-07T07:36:03.8314182Z * [new branch] gh/colinchan15/3/head -> origin/gh/colinchan15/3/head 2025-09-07T07:36:03.8315815Z * [new branch] gh/colinchan15/6/base -> origin/gh/colinchan15/6/base 2025-09-07T07:36:03.8317005Z * [new branch] gh/colinchan15/6/head -> origin/gh/colinchan15/6/head 2025-09-07T07:36:03.8319321Z * [new branch] gh/davidberard98/382/base -> origin/gh/davidberard98/382/base 2025-09-07T07:36:03.8320734Z * [new branch] gh/davidberard98/382/head -> origin/gh/davidberard98/382/head 2025-09-07T07:36:03.8321936Z * [new branch] gh/davidberard98/382/orig -> origin/gh/davidberard98/382/orig 2025-09-07T07:36:03.8323663Z * [new branch] gh/davidberard98/386/base -> origin/gh/davidberard98/386/base 2025-09-07T07:36:03.8324892Z * [new branch] gh/davidberard98/386/head -> origin/gh/davidberard98/386/head 2025-09-07T07:36:03.8326175Z * [new branch] gh/davidberard98/386/orig -> origin/gh/davidberard98/386/orig 2025-09-07T07:36:03.8327936Z * [new branch] gh/davidberard98/391/base -> origin/gh/davidberard98/391/base 2025-09-07T07:36:03.8329141Z * [new branch] gh/davidberard98/391/head -> origin/gh/davidberard98/391/head 2025-09-07T07:36:03.8330238Z * [new branch] gh/davidberard98/391/orig -> origin/gh/davidberard98/391/orig 2025-09-07T07:36:03.8332335Z * [new branch] gh/davidberard98/392/base -> origin/gh/davidberard98/392/base 2025-09-07T07:36:03.8333132Z * [new branch] gh/davidberard98/392/head -> origin/gh/davidberard98/392/head 2025-09-07T07:36:03.8334328Z * [new branch] gh/davidberard98/392/orig -> origin/gh/davidberard98/392/orig 2025-09-07T07:36:03.8336366Z * [new branch] gh/davidberard98/394/base -> origin/gh/davidberard98/394/base 2025-09-07T07:36:03.8337499Z * [new branch] gh/davidberard98/394/head -> origin/gh/davidberard98/394/head 2025-09-07T07:36:03.8338814Z * [new branch] gh/davidberard98/394/orig -> origin/gh/davidberard98/394/orig 2025-09-07T07:36:03.8340541Z * [new branch] gh/davidberard98/396/base -> origin/gh/davidberard98/396/base 2025-09-07T07:36:03.8341714Z * [new branch] gh/davidberard98/396/head -> origin/gh/davidberard98/396/head 2025-09-07T07:36:03.8343129Z * [new branch] gh/davidberard98/396/orig -> origin/gh/davidberard98/396/orig 2025-09-07T07:36:03.8345022Z * [new branch] gh/davidberard98/397/base -> origin/gh/davidberard98/397/base 2025-09-07T07:36:03.8346226Z * [new branch] gh/davidberard98/397/head -> origin/gh/davidberard98/397/head 2025-09-07T07:36:03.8347495Z * [new branch] gh/davidberard98/397/orig -> origin/gh/davidberard98/397/orig 2025-09-07T07:36:03.8349231Z * [new branch] gh/davidberard98/398/base -> origin/gh/davidberard98/398/base 2025-09-07T07:36:03.8350397Z * [new branch] gh/davidberard98/398/head -> origin/gh/davidberard98/398/head 2025-09-07T07:36:03.8351647Z * [new branch] gh/davidberard98/398/orig -> origin/gh/davidberard98/398/orig 2025-09-07T07:36:03.8353505Z * [new branch] gh/davidberard98/399/base -> origin/gh/davidberard98/399/base 2025-09-07T07:36:03.8354739Z * [new branch] gh/davidberard98/399/head -> origin/gh/davidberard98/399/head 2025-09-07T07:36:03.8355946Z * [new branch] gh/davidberard98/399/orig -> origin/gh/davidberard98/399/orig 2025-09-07T07:36:03.8357687Z * [new branch] gh/davidberard98/400/base -> origin/gh/davidberard98/400/base 2025-09-07T07:36:03.8358955Z * [new branch] gh/davidberard98/400/head -> origin/gh/davidberard98/400/head 2025-09-07T07:36:03.8360162Z * [new branch] gh/davidberard98/400/orig -> origin/gh/davidberard98/400/orig 2025-09-07T07:36:03.8361843Z * [new branch] gh/davidberard98/401/base -> origin/gh/davidberard98/401/base 2025-09-07T07:36:03.8363076Z * [new branch] gh/davidberard98/401/head -> origin/gh/davidberard98/401/head 2025-09-07T07:36:03.8364337Z * [new branch] gh/davidberard98/401/orig -> origin/gh/davidberard98/401/orig 2025-09-07T07:36:03.8366045Z * [new branch] gh/davidberard98/402/base -> origin/gh/davidberard98/402/base 2025-09-07T07:36:03.8367249Z * [new branch] gh/davidberard98/402/head -> origin/gh/davidberard98/402/head 2025-09-07T07:36:03.8368445Z * [new branch] gh/davidberard98/402/orig -> origin/gh/davidberard98/402/orig 2025-09-07T07:36:03.8370235Z * [new branch] gh/davidberard98/403/base -> origin/gh/davidberard98/403/base 2025-09-07T07:36:03.8371434Z * [new branch] gh/davidberard98/403/head -> origin/gh/davidberard98/403/head 2025-09-07T07:36:03.8372624Z * [new branch] gh/davidberard98/403/orig -> origin/gh/davidberard98/403/orig 2025-09-07T07:36:03.8374490Z * [new branch] gh/davidberard98/404/base -> origin/gh/davidberard98/404/base 2025-09-07T07:36:03.8375688Z * [new branch] gh/davidberard98/404/head -> origin/gh/davidberard98/404/head 2025-09-07T07:36:03.8376784Z * [new branch] gh/davidberard98/404/orig -> origin/gh/davidberard98/404/orig 2025-09-07T07:36:03.8378700Z * [new branch] gh/davidberard98/405/base -> origin/gh/davidberard98/405/base 2025-09-07T07:36:03.8379921Z * [new branch] gh/davidberard98/405/head -> origin/gh/davidberard98/405/head 2025-09-07T07:36:03.8381097Z * [new branch] gh/davidberard98/405/orig -> origin/gh/davidberard98/405/orig 2025-09-07T07:36:03.8382940Z * [new branch] gh/davidberard98/406/base -> origin/gh/davidberard98/406/base 2025-09-07T07:36:03.8384226Z * [new branch] gh/davidberard98/406/head -> origin/gh/davidberard98/406/head 2025-09-07T07:36:03.8385500Z * [new branch] gh/davidberard98/406/orig -> origin/gh/davidberard98/406/orig 2025-09-07T07:36:03.8387931Z * [new branch] gh/davidberard98/407/base -> origin/gh/davidberard98/407/base 2025-09-07T07:36:03.8389126Z * [new branch] gh/davidberard98/407/head -> origin/gh/davidberard98/407/head 2025-09-07T07:36:03.8390375Z * [new branch] gh/davidberard98/407/orig -> origin/gh/davidberard98/407/orig 2025-09-07T07:36:03.8392161Z * [new branch] gh/davidberard98/408/base -> origin/gh/davidberard98/408/base 2025-09-07T07:36:03.8393332Z * [new branch] gh/davidberard98/408/head -> origin/gh/davidberard98/408/head 2025-09-07T07:36:03.8394494Z * [new branch] gh/davidberard98/408/orig -> origin/gh/davidberard98/408/orig 2025-09-07T07:36:03.8396163Z * [new branch] gh/davidberard98/409/base -> origin/gh/davidberard98/409/base 2025-09-07T07:36:03.8397423Z * [new branch] gh/davidberard98/409/head -> origin/gh/davidberard98/409/head 2025-09-07T07:36:03.8398712Z * [new branch] gh/davidberard98/409/orig -> origin/gh/davidberard98/409/orig 2025-09-07T07:36:03.8400888Z * [new branch] gh/desertfire/594/base -> origin/gh/desertfire/594/base 2025-09-07T07:36:03.8402078Z * [new branch] gh/desertfire/594/head -> origin/gh/desertfire/594/head 2025-09-07T07:36:03.8403439Z * [new branch] gh/desertfire/594/orig -> origin/gh/desertfire/594/orig 2025-09-07T07:36:03.8405308Z * [new branch] gh/desertfire/595/base -> origin/gh/desertfire/595/base 2025-09-07T07:36:03.8406443Z * [new branch] gh/desertfire/595/head -> origin/gh/desertfire/595/head 2025-09-07T07:36:03.8407656Z * [new branch] gh/desertfire/595/orig -> origin/gh/desertfire/595/orig 2025-09-07T07:36:03.8409467Z * [new branch] gh/desertfire/597/base -> origin/gh/desertfire/597/base 2025-09-07T07:36:03.8410646Z * [new branch] gh/desertfire/597/head -> origin/gh/desertfire/597/head 2025-09-07T07:36:03.8411888Z * [new branch] gh/desertfire/597/orig -> origin/gh/desertfire/597/orig 2025-09-07T07:36:03.8414089Z * [new branch] gh/dharakk/1/base -> origin/gh/dharakk/1/base 2025-09-07T07:36:03.8415443Z * [new branch] gh/dharakk/1/head -> origin/gh/dharakk/1/head 2025-09-07T07:36:03.8417633Z * [new branch] gh/drisspg/149/base -> origin/gh/drisspg/149/base 2025-09-07T07:36:03.8418866Z * [new branch] gh/drisspg/149/head -> origin/gh/drisspg/149/head 2025-09-07T07:36:03.8420144Z * [new branch] gh/drisspg/149/orig -> origin/gh/drisspg/149/orig 2025-09-07T07:36:03.8421833Z * [new branch] gh/drisspg/159/base -> origin/gh/drisspg/159/base 2025-09-07T07:36:03.8422999Z * [new branch] gh/drisspg/159/head -> origin/gh/drisspg/159/head 2025-09-07T07:36:03.8424187Z * [new branch] gh/drisspg/159/orig -> origin/gh/drisspg/159/orig 2025-09-07T07:36:03.8426011Z * [new branch] gh/drisspg/166/base -> origin/gh/drisspg/166/base 2025-09-07T07:36:03.8427111Z * [new branch] gh/drisspg/166/head -> origin/gh/drisspg/166/head 2025-09-07T07:36:03.8428320Z * [new branch] gh/drisspg/166/orig -> origin/gh/drisspg/166/orig 2025-09-07T07:36:03.8430019Z * [new branch] gh/drisspg/170/base -> origin/gh/drisspg/170/base 2025-09-07T07:36:03.8431189Z * [new branch] gh/drisspg/170/head -> origin/gh/drisspg/170/head 2025-09-07T07:36:03.8432413Z * [new branch] gh/drisspg/170/orig -> origin/gh/drisspg/170/orig 2025-09-07T07:36:03.8434105Z * [new branch] gh/drisspg/173/base -> origin/gh/drisspg/173/base 2025-09-07T07:36:03.8435291Z * [new branch] gh/drisspg/173/head -> origin/gh/drisspg/173/head 2025-09-07T07:36:03.8436565Z * [new branch] gh/drisspg/173/orig -> origin/gh/drisspg/173/orig 2025-09-07T07:36:03.8438266Z * [new branch] gh/drisspg/177/base -> origin/gh/drisspg/177/base 2025-09-07T07:36:03.8439463Z * [new branch] gh/drisspg/177/head -> origin/gh/drisspg/177/head 2025-09-07T07:36:03.8440625Z * [new branch] gh/drisspg/177/orig -> origin/gh/drisspg/177/orig 2025-09-07T07:36:03.8442316Z * [new branch] gh/drisspg/178/base -> origin/gh/drisspg/178/base 2025-09-07T07:36:03.8443497Z * [new branch] gh/drisspg/178/head -> origin/gh/drisspg/178/head 2025-09-07T07:36:03.8444586Z * [new branch] gh/drisspg/178/orig -> origin/gh/drisspg/178/orig 2025-09-07T07:36:03.8446302Z * [new branch] gh/drisspg/180/base -> origin/gh/drisspg/180/base 2025-09-07T07:36:03.8447580Z * [new branch] gh/drisspg/180/head -> origin/gh/drisspg/180/head 2025-09-07T07:36:03.8448740Z * [new branch] gh/drisspg/180/orig -> origin/gh/drisspg/180/orig 2025-09-07T07:36:03.8450532Z * [new branch] gh/drisspg/181/base -> origin/gh/drisspg/181/base 2025-09-07T07:36:03.8451773Z * [new branch] gh/drisspg/181/head -> origin/gh/drisspg/181/head 2025-09-07T07:36:03.8453113Z * [new branch] gh/drisspg/181/orig -> origin/gh/drisspg/181/orig 2025-09-07T07:36:03.8454822Z * [new branch] gh/drisspg/182/base -> origin/gh/drisspg/182/base 2025-09-07T07:36:03.8456039Z * [new branch] gh/drisspg/182/head -> origin/gh/drisspg/182/head 2025-09-07T07:36:03.8457723Z * [new branch] gh/drisspg/183/base -> origin/gh/drisspg/183/base 2025-09-07T07:36:03.8458815Z * [new branch] gh/drisspg/183/head -> origin/gh/drisspg/183/head 2025-09-07T07:36:03.8460398Z * [new branch] gh/drisspg/184/base -> origin/gh/drisspg/184/base 2025-09-07T07:36:03.8461582Z * [new branch] gh/drisspg/184/head -> origin/gh/drisspg/184/head 2025-09-07T07:36:03.8463318Z * [new branch] gh/drisspg/185/base -> origin/gh/drisspg/185/base 2025-09-07T07:36:03.8464499Z * [new branch] gh/drisspg/185/head -> origin/gh/drisspg/185/head 2025-09-07T07:36:03.8466180Z * [new branch] gh/drisspg/186/base -> origin/gh/drisspg/186/base 2025-09-07T07:36:03.8467385Z * [new branch] gh/drisspg/186/head -> origin/gh/drisspg/186/head 2025-09-07T07:36:03.8468644Z * [new branch] gh/drisspg/186/orig -> origin/gh/drisspg/186/orig 2025-09-07T07:36:03.8470375Z * [new branch] gh/drisspg/187/base -> origin/gh/drisspg/187/base 2025-09-07T07:36:03.8471548Z * [new branch] gh/drisspg/187/head -> origin/gh/drisspg/187/head 2025-09-07T07:36:03.8472742Z * [new branch] gh/drisspg/187/orig -> origin/gh/drisspg/187/orig 2025-09-07T07:36:03.8474496Z * [new branch] gh/drisspg/188/base -> origin/gh/drisspg/188/base 2025-09-07T07:36:03.8475750Z * [new branch] gh/drisspg/188/head -> origin/gh/drisspg/188/head 2025-09-07T07:36:03.8476885Z * [new branch] gh/drisspg/188/orig -> origin/gh/drisspg/188/orig 2025-09-07T07:36:03.8479023Z * [new branch] gh/drisspg/189/base -> origin/gh/drisspg/189/base 2025-09-07T07:36:03.8480291Z * [new branch] gh/drisspg/189/head -> origin/gh/drisspg/189/head 2025-09-07T07:36:03.8481522Z * [new branch] gh/drisspg/189/orig -> origin/gh/drisspg/189/orig 2025-09-07T07:36:03.8483274Z * [new branch] gh/drisspg/190/base -> origin/gh/drisspg/190/base 2025-09-07T07:36:03.8484497Z * [new branch] gh/drisspg/190/head -> origin/gh/drisspg/190/head 2025-09-07T07:36:03.8485900Z * [new branch] gh/drisspg/190/orig -> origin/gh/drisspg/190/orig 2025-09-07T07:36:03.8487643Z * [new branch] gh/drisspg/191/base -> origin/gh/drisspg/191/base 2025-09-07T07:36:03.8488874Z * [new branch] gh/drisspg/191/head -> origin/gh/drisspg/191/head 2025-09-07T07:36:03.8490096Z * [new branch] gh/drisspg/191/orig -> origin/gh/drisspg/191/orig 2025-09-07T07:36:03.8491856Z * [new branch] gh/drisspg/192/base -> origin/gh/drisspg/192/base 2025-09-07T07:36:03.8492967Z * [new branch] gh/drisspg/192/head -> origin/gh/drisspg/192/head 2025-09-07T07:36:03.8494160Z * [new branch] gh/drisspg/192/orig -> origin/gh/drisspg/192/orig 2025-09-07T07:36:03.8495903Z * [new branch] gh/drisspg/193/base -> origin/gh/drisspg/193/base 2025-09-07T07:36:03.8497194Z * [new branch] gh/drisspg/193/head -> origin/gh/drisspg/193/head 2025-09-07T07:36:03.8498479Z * [new branch] gh/drisspg/193/orig -> origin/gh/drisspg/193/orig 2025-09-07T07:36:03.8500175Z * [new branch] gh/drisspg/194/base -> origin/gh/drisspg/194/base 2025-09-07T07:36:03.8501378Z * [new branch] gh/drisspg/194/head -> origin/gh/drisspg/194/head 2025-09-07T07:36:03.8502629Z * [new branch] gh/drisspg/194/orig -> origin/gh/drisspg/194/orig 2025-09-07T07:36:03.8508361Z * [new branch] gh/drisspg/195/base -> origin/gh/drisspg/195/base 2025-09-07T07:36:03.8508909Z * [new branch] gh/drisspg/195/head -> origin/gh/drisspg/195/head 2025-09-07T07:36:03.8510252Z * [new branch] gh/drisspg/195/orig -> origin/gh/drisspg/195/orig 2025-09-07T07:36:03.8512270Z * [new branch] gh/drisspg/196/base -> origin/gh/drisspg/196/base 2025-09-07T07:36:03.8513571Z * [new branch] gh/drisspg/196/head -> origin/gh/drisspg/196/head 2025-09-07T07:36:03.8514779Z * [new branch] gh/drisspg/196/orig -> origin/gh/drisspg/196/orig 2025-09-07T07:36:03.8516627Z * [new branch] gh/drisspg/197/base -> origin/gh/drisspg/197/base 2025-09-07T07:36:03.8518156Z * [new branch] gh/drisspg/197/head -> origin/gh/drisspg/197/head 2025-09-07T07:36:03.8519363Z * [new branch] gh/drisspg/197/orig -> origin/gh/drisspg/197/orig 2025-09-07T07:36:03.8521333Z * [new branch] gh/drisspg/198/base -> origin/gh/drisspg/198/base 2025-09-07T07:36:03.8522552Z * [new branch] gh/drisspg/198/head -> origin/gh/drisspg/198/head 2025-09-07T07:36:03.8523785Z * [new branch] gh/drisspg/198/orig -> origin/gh/drisspg/198/orig 2025-09-07T07:36:03.8525745Z * [new branch] gh/drisspg/199/base -> origin/gh/drisspg/199/base 2025-09-07T07:36:03.8526911Z * [new branch] gh/drisspg/199/head -> origin/gh/drisspg/199/head 2025-09-07T07:36:03.8528246Z * [new branch] gh/drisspg/199/orig -> origin/gh/drisspg/199/orig 2025-09-07T07:36:03.8530752Z * [new branch] gh/dsjohns2/1/base -> origin/gh/dsjohns2/1/base 2025-09-07T07:36:03.8531839Z * [new branch] gh/dsjohns2/1/head -> origin/gh/dsjohns2/1/head 2025-09-07T07:36:03.8534254Z * [new branch] gh/eellison/784/base -> origin/gh/eellison/784/base 2025-09-07T07:36:03.8535453Z * [new branch] gh/eellison/784/head -> origin/gh/eellison/784/head 2025-09-07T07:36:03.8536646Z * [new branch] gh/eellison/784/orig -> origin/gh/eellison/784/orig 2025-09-07T07:36:03.8538867Z * [new branch] gh/eellison/785/base -> origin/gh/eellison/785/base 2025-09-07T07:36:03.8540068Z * [new branch] gh/eellison/785/head -> origin/gh/eellison/785/head 2025-09-07T07:36:03.8541289Z * [new branch] gh/eellison/785/orig -> origin/gh/eellison/785/orig 2025-09-07T07:36:03.8543061Z * [new branch] gh/eellison/789/base -> origin/gh/eellison/789/base 2025-09-07T07:36:03.8544254Z * [new branch] gh/eellison/789/head -> origin/gh/eellison/789/head 2025-09-07T07:36:03.8545429Z * [new branch] gh/eellison/789/orig -> origin/gh/eellison/789/orig 2025-09-07T07:36:03.8547792Z * [new branch] gh/eellison/800/base -> origin/gh/eellison/800/base 2025-09-07T07:36:03.8548994Z * [new branch] gh/eellison/800/head -> origin/gh/eellison/800/head 2025-09-07T07:36:03.8550252Z * [new branch] gh/eellison/800/orig -> origin/gh/eellison/800/orig 2025-09-07T07:36:03.8552014Z * [new branch] gh/eellison/801/base -> origin/gh/eellison/801/base 2025-09-07T07:36:03.8553210Z * [new branch] gh/eellison/801/head -> origin/gh/eellison/801/head 2025-09-07T07:36:03.8554426Z * [new branch] gh/eellison/801/orig -> origin/gh/eellison/801/orig 2025-09-07T07:36:03.8556321Z * [new branch] gh/eellison/802/base -> origin/gh/eellison/802/base 2025-09-07T07:36:03.8557716Z * [new branch] gh/eellison/802/head -> origin/gh/eellison/802/head 2025-09-07T07:36:03.8558932Z * [new branch] gh/eellison/802/orig -> origin/gh/eellison/802/orig 2025-09-07T07:36:03.8560686Z * [new branch] gh/eellison/805/base -> origin/gh/eellison/805/base 2025-09-07T07:36:03.8561851Z * [new branch] gh/eellison/805/head -> origin/gh/eellison/805/head 2025-09-07T07:36:03.8563104Z * [new branch] gh/eellison/805/orig -> origin/gh/eellison/805/orig 2025-09-07T07:36:03.8565029Z * [new branch] gh/eellison/808/base -> origin/gh/eellison/808/base 2025-09-07T07:36:03.8566291Z * [new branch] gh/eellison/808/head -> origin/gh/eellison/808/head 2025-09-07T07:36:03.8567496Z * [new branch] gh/eellison/808/orig -> origin/gh/eellison/808/orig 2025-09-07T07:36:03.8569243Z * [new branch] gh/eellison/809/base -> origin/gh/eellison/809/base 2025-09-07T07:36:03.8570439Z * [new branch] gh/eellison/809/head -> origin/gh/eellison/809/head 2025-09-07T07:36:03.8571612Z * [new branch] gh/eellison/809/orig -> origin/gh/eellison/809/orig 2025-09-07T07:36:03.8573500Z * [new branch] gh/eellison/813/base -> origin/gh/eellison/813/base 2025-09-07T07:36:03.8574743Z * [new branch] gh/eellison/813/head -> origin/gh/eellison/813/head 2025-09-07T07:36:03.8575936Z * [new branch] gh/eellison/813/orig -> origin/gh/eellison/813/orig 2025-09-07T07:36:03.8578201Z * [new branch] gh/eellison/814/base -> origin/gh/eellison/814/base 2025-09-07T07:36:03.8579041Z * [new branch] gh/eellison/814/head -> origin/gh/eellison/814/head 2025-09-07T07:36:03.8580264Z * [new branch] gh/eellison/814/orig -> origin/gh/eellison/814/orig 2025-09-07T07:36:03.8582730Z * [new branch] gh/eellison/815/base -> origin/gh/eellison/815/base 2025-09-07T07:36:03.8583656Z * [new branch] gh/eellison/815/head -> origin/gh/eellison/815/head 2025-09-07T07:36:03.8584951Z * [new branch] gh/eellison/815/orig -> origin/gh/eellison/815/orig 2025-09-07T07:36:03.8586680Z * [new branch] gh/eellison/816/base -> origin/gh/eellison/816/base 2025-09-07T07:36:03.8587845Z * [new branch] gh/eellison/816/head -> origin/gh/eellison/816/head 2025-09-07T07:36:03.8589050Z * [new branch] gh/eellison/816/orig -> origin/gh/eellison/816/orig 2025-09-07T07:36:03.8590891Z * [new branch] gh/eellison/817/base -> origin/gh/eellison/817/base 2025-09-07T07:36:03.8592107Z * [new branch] gh/eellison/817/head -> origin/gh/eellison/817/head 2025-09-07T07:36:03.8593379Z * [new branch] gh/eellison/817/orig -> origin/gh/eellison/817/orig 2025-09-07T07:36:03.8595277Z * [new branch] gh/eellison/818/base -> origin/gh/eellison/818/base 2025-09-07T07:36:03.8596400Z * [new branch] gh/eellison/818/head -> origin/gh/eellison/818/head 2025-09-07T07:36:03.8597583Z * [new branch] gh/eellison/818/orig -> origin/gh/eellison/818/orig 2025-09-07T07:36:03.8599539Z * [new branch] gh/eellison/819/base -> origin/gh/eellison/819/base 2025-09-07T07:36:03.8600683Z * [new branch] gh/eellison/819/head -> origin/gh/eellison/819/head 2025-09-07T07:36:03.8601855Z * [new branch] gh/eellison/819/orig -> origin/gh/eellison/819/orig 2025-09-07T07:36:03.8604501Z * [new branch] gh/eellison/820/base -> origin/gh/eellison/820/base 2025-09-07T07:36:03.8605807Z * [new branch] gh/eellison/820/head -> origin/gh/eellison/820/head 2025-09-07T07:36:03.8607014Z * [new branch] gh/eellison/820/orig -> origin/gh/eellison/820/orig 2025-09-07T07:36:03.8608858Z * [new branch] gh/eellison/821/base -> origin/gh/eellison/821/base 2025-09-07T07:36:03.8610084Z * [new branch] gh/eellison/821/head -> origin/gh/eellison/821/head 2025-09-07T07:36:03.8611304Z * [new branch] gh/eellison/821/orig -> origin/gh/eellison/821/orig 2025-09-07T07:36:03.8613055Z * [new branch] gh/eellison/822/base -> origin/gh/eellison/822/base 2025-09-07T07:36:03.8614232Z * [new branch] gh/eellison/822/head -> origin/gh/eellison/822/head 2025-09-07T07:36:03.8615449Z * [new branch] gh/eellison/822/orig -> origin/gh/eellison/822/orig 2025-09-07T07:36:03.8617313Z * [new branch] gh/eellison/823/base -> origin/gh/eellison/823/base 2025-09-07T07:36:03.8618587Z * [new branch] gh/eellison/823/head -> origin/gh/eellison/823/head 2025-09-07T07:36:03.8619876Z * [new branch] gh/eellison/823/orig -> origin/gh/eellison/823/orig 2025-09-07T07:36:03.8622027Z * [new branch] gh/etaf/132/base -> origin/gh/etaf/132/base 2025-09-07T07:36:03.8623231Z * [new branch] gh/etaf/132/head -> origin/gh/etaf/132/head 2025-09-07T07:36:03.8624514Z * [new branch] gh/etaf/132/orig -> origin/gh/etaf/132/orig 2025-09-07T07:36:03.8626222Z * [new branch] gh/etaf/138/base -> origin/gh/etaf/138/base 2025-09-07T07:36:03.8627405Z * [new branch] gh/etaf/138/head -> origin/gh/etaf/138/head 2025-09-07T07:36:03.8629021Z * [new branch] gh/etaf/138/orig -> origin/gh/etaf/138/orig 2025-09-07T07:36:03.8630936Z * [new branch] gh/etaf/140/base -> origin/gh/etaf/140/base 2025-09-07T07:36:03.8632108Z * [new branch] gh/etaf/140/head -> origin/gh/etaf/140/head 2025-09-07T07:36:03.8633432Z * [new branch] gh/etaf/140/orig -> origin/gh/etaf/140/orig 2025-09-07T07:36:03.8635066Z * [new branch] gh/etaf/143/base -> origin/gh/etaf/143/base 2025-09-07T07:36:03.8636272Z * [new branch] gh/etaf/143/head -> origin/gh/etaf/143/head 2025-09-07T07:36:03.8637470Z * [new branch] gh/etaf/143/orig -> origin/gh/etaf/143/orig 2025-09-07T07:36:03.8639180Z * [new branch] gh/etaf/147/base -> origin/gh/etaf/147/base 2025-09-07T07:36:03.8640399Z * [new branch] gh/etaf/147/head -> origin/gh/etaf/147/head 2025-09-07T07:36:03.8642250Z * [new branch] gh/etaf/151/base -> origin/gh/etaf/151/base 2025-09-07T07:36:03.8643532Z * [new branch] gh/etaf/151/head -> origin/gh/etaf/151/head 2025-09-07T07:36:03.8644745Z * [new branch] gh/etaf/151/orig -> origin/gh/etaf/151/orig 2025-09-07T07:36:03.8646524Z * [new branch] gh/etaf/152/base -> origin/gh/etaf/152/base 2025-09-07T07:36:03.8648234Z * [new branch] gh/etaf/152/head -> origin/gh/etaf/152/head 2025-09-07T07:36:03.8649437Z * [new branch] gh/etaf/152/orig -> origin/gh/etaf/152/orig 2025-09-07T07:36:03.8651686Z * [new branch] gh/etaf/153/base -> origin/gh/etaf/153/base 2025-09-07T07:36:03.8652935Z * [new branch] gh/etaf/153/head -> origin/gh/etaf/153/head 2025-09-07T07:36:03.8654137Z * [new branch] gh/etaf/153/orig -> origin/gh/etaf/153/orig 2025-09-07T07:36:03.8656021Z * [new branch] gh/etaf/154/base -> origin/gh/etaf/154/base 2025-09-07T07:36:03.8657250Z * [new branch] gh/etaf/154/head -> origin/gh/etaf/154/head 2025-09-07T07:36:03.8658561Z * [new branch] gh/etaf/154/orig -> origin/gh/etaf/154/orig 2025-09-07T07:36:03.8660399Z * [new branch] gh/etaf/155/base -> origin/gh/etaf/155/base 2025-09-07T07:36:03.8661679Z * [new branch] gh/etaf/155/head -> origin/gh/etaf/155/head 2025-09-07T07:36:03.8662848Z * [new branch] gh/etaf/155/orig -> origin/gh/etaf/155/orig 2025-09-07T07:36:03.8664500Z * [new branch] gh/etaf/156/base -> origin/gh/etaf/156/base 2025-09-07T07:36:03.8665800Z * [new branch] gh/etaf/156/head -> origin/gh/etaf/156/head 2025-09-07T07:36:03.8667139Z * [new branch] gh/etaf/156/orig -> origin/gh/etaf/156/orig 2025-09-07T07:36:03.8668977Z * [new branch] gh/etaf/157/base -> origin/gh/etaf/157/base 2025-09-07T07:36:03.8670192Z * [new branch] gh/etaf/157/head -> origin/gh/etaf/157/head 2025-09-07T07:36:03.8671409Z * [new branch] gh/etaf/157/orig -> origin/gh/etaf/157/orig 2025-09-07T07:36:03.8673019Z * [new branch] gh/etaf/158/base -> origin/gh/etaf/158/base 2025-09-07T07:36:03.8674259Z * [new branch] gh/etaf/158/head -> origin/gh/etaf/158/head 2025-09-07T07:36:03.8675491Z * [new branch] gh/etaf/158/orig -> origin/gh/etaf/158/orig 2025-09-07T07:36:03.8677310Z * [new branch] gh/etaf/159/base -> origin/gh/etaf/159/base 2025-09-07T07:36:03.8678570Z * [new branch] gh/etaf/159/head -> origin/gh/etaf/159/head 2025-09-07T07:36:03.8679742Z * [new branch] gh/etaf/159/orig -> origin/gh/etaf/159/orig 2025-09-07T07:36:03.8681933Z * [new branch] gh/etaf/160/base -> origin/gh/etaf/160/base 2025-09-07T07:36:03.8683186Z * [new branch] gh/etaf/160/head -> origin/gh/etaf/160/head 2025-09-07T07:36:03.8684418Z * [new branch] gh/etaf/160/orig -> origin/gh/etaf/160/orig 2025-09-07T07:36:03.8686210Z * [new branch] gh/etaf/161/base -> origin/gh/etaf/161/base 2025-09-07T07:36:03.8687488Z * [new branch] gh/etaf/161/head -> origin/gh/etaf/161/head 2025-09-07T07:36:03.8688646Z * [new branch] gh/etaf/161/orig -> origin/gh/etaf/161/orig 2025-09-07T07:36:03.8690800Z * [new branch] gh/etaf/162/base -> origin/gh/etaf/162/base 2025-09-07T07:36:03.8692027Z * [new branch] gh/etaf/162/head -> origin/gh/etaf/162/head 2025-09-07T07:36:03.8693227Z * [new branch] gh/etaf/162/orig -> origin/gh/etaf/162/orig 2025-09-07T07:36:03.8695059Z * [new branch] gh/etaf/163/base -> origin/gh/etaf/163/base 2025-09-07T07:36:03.8696256Z * [new branch] gh/etaf/163/head -> origin/gh/etaf/163/head 2025-09-07T07:36:03.8697418Z * [new branch] gh/etaf/163/orig -> origin/gh/etaf/163/orig 2025-09-07T07:36:03.8699345Z * [new branch] gh/etaf/164/base -> origin/gh/etaf/164/base 2025-09-07T07:36:03.8700603Z * [new branch] gh/etaf/164/head -> origin/gh/etaf/164/head 2025-09-07T07:36:03.8702206Z * [new branch] gh/etaf/164/orig -> origin/gh/etaf/164/orig 2025-09-07T07:36:03.8704428Z * [new branch] gh/etaf/165/base -> origin/gh/etaf/165/base 2025-09-07T07:36:03.8705624Z * [new branch] gh/etaf/165/orig -> origin/gh/etaf/165/orig 2025-09-07T07:36:03.8707875Z * [new branch] gh/etaf/166/base -> origin/gh/etaf/166/base 2025-09-07T07:36:03.8709096Z * [new branch] gh/etaf/166/head -> origin/gh/etaf/166/head 2025-09-07T07:36:03.8710305Z * [new branch] gh/etaf/166/orig -> origin/gh/etaf/166/orig 2025-09-07T07:36:03.8712232Z * [new branch] gh/etaf/167/base -> origin/gh/etaf/167/base 2025-09-07T07:36:03.8713503Z * [new branch] gh/etaf/167/head -> origin/gh/etaf/167/head 2025-09-07T07:36:03.8714704Z * [new branch] gh/etaf/167/orig -> origin/gh/etaf/167/orig 2025-09-07T07:36:03.8716472Z * [new branch] gh/etaf/168/base -> origin/gh/etaf/168/base 2025-09-07T07:36:03.8717735Z * [new branch] gh/etaf/168/head -> origin/gh/etaf/168/head 2025-09-07T07:36:03.8718968Z * [new branch] gh/etaf/168/orig -> origin/gh/etaf/168/orig 2025-09-07T07:36:03.8720841Z * [new branch] gh/etaf/169/base -> origin/gh/etaf/169/base 2025-09-07T07:36:03.8722041Z * [new branch] gh/etaf/169/head -> origin/gh/etaf/169/head 2025-09-07T07:36:03.8723433Z * [new branch] gh/etaf/169/orig -> origin/gh/etaf/169/orig 2025-09-07T07:36:03.8725423Z * [new branch] gh/exclamaforte/1/base -> origin/gh/exclamaforte/1/base 2025-09-07T07:36:03.8726632Z * [new branch] gh/exclamaforte/1/head -> origin/gh/exclamaforte/1/head 2025-09-07T07:36:03.8728320Z * [new branch] gh/exclamaforte/2/base -> origin/gh/exclamaforte/2/base 2025-09-07T07:36:03.8729451Z * [new branch] gh/exclamaforte/2/head -> origin/gh/exclamaforte/2/head 2025-09-07T07:36:03.8731356Z * [new branch] gh/exclamaforte/3/base -> origin/gh/exclamaforte/3/base 2025-09-07T07:36:03.8732398Z * [new branch] gh/exclamaforte/3/head -> origin/gh/exclamaforte/3/head 2025-09-07T07:36:03.8734054Z * [new branch] gh/exclamaforte/4/base -> origin/gh/exclamaforte/4/base 2025-09-07T07:36:03.8735206Z * [new branch] gh/exclamaforte/4/head -> origin/gh/exclamaforte/4/head 2025-09-07T07:36:03.8737573Z * [new branch] gh/ezyang/2374/base -> origin/gh/ezyang/2374/base 2025-09-07T07:36:03.8738940Z * [new branch] gh/ezyang/2374/head -> origin/gh/ezyang/2374/head 2025-09-07T07:36:03.8740280Z * [new branch] gh/ezyang/2374/orig -> origin/gh/ezyang/2374/orig 2025-09-07T07:36:03.8742072Z * [new branch] gh/ezyang/2973/base -> origin/gh/ezyang/2973/base 2025-09-07T07:36:03.8743395Z * [new branch] gh/ezyang/2973/head -> origin/gh/ezyang/2973/head 2025-09-07T07:36:03.8744711Z * [new branch] gh/ezyang/2973/orig -> origin/gh/ezyang/2973/orig 2025-09-07T07:36:03.8746578Z * [new branch] gh/ezyang/2974/base -> origin/gh/ezyang/2974/base 2025-09-07T07:36:03.8747782Z * [new branch] gh/ezyang/2974/head -> origin/gh/ezyang/2974/head 2025-09-07T07:36:03.8749034Z * [new branch] gh/ezyang/2974/orig -> origin/gh/ezyang/2974/orig 2025-09-07T07:36:03.8750780Z * [new branch] gh/ezyang/3074/base -> origin/gh/ezyang/3074/base 2025-09-07T07:36:03.8751936Z * [new branch] gh/ezyang/3074/head -> origin/gh/ezyang/3074/head 2025-09-07T07:36:03.8753179Z * [new branch] gh/ezyang/3074/orig -> origin/gh/ezyang/3074/orig 2025-09-07T07:36:03.8754858Z * [new branch] gh/ezyang/3088/base -> origin/gh/ezyang/3088/base 2025-09-07T07:36:03.8756036Z * [new branch] gh/ezyang/3088/head -> origin/gh/ezyang/3088/head 2025-09-07T07:36:03.8757338Z * [new branch] gh/ezyang/3088/orig -> origin/gh/ezyang/3088/orig 2025-09-07T07:36:03.8759020Z * [new branch] gh/ezyang/3092/base -> origin/gh/ezyang/3092/base 2025-09-07T07:36:03.8760706Z * [new branch] gh/ezyang/3092/head -> origin/gh/ezyang/3092/head 2025-09-07T07:36:03.8761462Z * [new branch] gh/ezyang/3092/orig -> origin/gh/ezyang/3092/orig 2025-09-07T07:36:03.8763198Z * [new branch] gh/ezyang/3103/base -> origin/gh/ezyang/3103/base 2025-09-07T07:36:03.8764385Z * [new branch] gh/ezyang/3103/head -> origin/gh/ezyang/3103/head 2025-09-07T07:36:03.8765555Z * [new branch] gh/ezyang/3103/orig -> origin/gh/ezyang/3103/orig 2025-09-07T07:36:03.8767254Z * [new branch] gh/ezyang/3105/base -> origin/gh/ezyang/3105/base 2025-09-07T07:36:03.8768417Z * [new branch] gh/ezyang/3105/head -> origin/gh/ezyang/3105/head 2025-09-07T07:36:03.8769610Z * [new branch] gh/ezyang/3105/orig -> origin/gh/ezyang/3105/orig 2025-09-07T07:36:03.8771365Z * [new branch] gh/ezyang/3114/base -> origin/gh/ezyang/3114/base 2025-09-07T07:36:03.8772607Z * [new branch] gh/ezyang/3114/head -> origin/gh/ezyang/3114/head 2025-09-07T07:36:03.8773827Z * [new branch] gh/ezyang/3114/orig -> origin/gh/ezyang/3114/orig 2025-09-07T07:36:03.8775480Z * [new branch] gh/ezyang/3116/base -> origin/gh/ezyang/3116/base 2025-09-07T07:36:03.8776633Z * [new branch] gh/ezyang/3116/head -> origin/gh/ezyang/3116/head 2025-09-07T07:36:03.8778135Z * [new branch] gh/ezyang/3116/orig -> origin/gh/ezyang/3116/orig 2025-09-07T07:36:03.8779827Z * [new branch] gh/ezyang/3120/base -> origin/gh/ezyang/3120/base 2025-09-07T07:36:03.8781007Z * [new branch] gh/ezyang/3120/head -> origin/gh/ezyang/3120/head 2025-09-07T07:36:03.8782174Z * [new branch] gh/ezyang/3120/orig -> origin/gh/ezyang/3120/orig 2025-09-07T07:36:03.8783938Z * [new branch] gh/ezyang/3122/base -> origin/gh/ezyang/3122/base 2025-09-07T07:36:03.8785170Z * [new branch] gh/ezyang/3122/head -> origin/gh/ezyang/3122/head 2025-09-07T07:36:03.8786400Z * [new branch] gh/ezyang/3122/orig -> origin/gh/ezyang/3122/orig 2025-09-07T07:36:03.8788115Z * [new branch] gh/ezyang/3123/base -> origin/gh/ezyang/3123/base 2025-09-07T07:36:03.8789319Z * [new branch] gh/ezyang/3123/head -> origin/gh/ezyang/3123/head 2025-09-07T07:36:03.8790586Z * [new branch] gh/ezyang/3123/orig -> origin/gh/ezyang/3123/orig 2025-09-07T07:36:03.8792208Z * [new branch] gh/ezyang/3125/base -> origin/gh/ezyang/3125/base 2025-09-07T07:36:03.8793423Z * [new branch] gh/ezyang/3125/head -> origin/gh/ezyang/3125/head 2025-09-07T07:36:03.8794679Z * [new branch] gh/ezyang/3125/orig -> origin/gh/ezyang/3125/orig 2025-09-07T07:36:03.8796332Z * [new branch] gh/ezyang/3126/base -> origin/gh/ezyang/3126/base 2025-09-07T07:36:03.8797499Z * [new branch] gh/ezyang/3126/head -> origin/gh/ezyang/3126/head 2025-09-07T07:36:03.8798659Z * [new branch] gh/ezyang/3126/orig -> origin/gh/ezyang/3126/orig 2025-09-07T07:36:03.8800847Z * [new branch] gh/ezyang/3127/base -> origin/gh/ezyang/3127/base 2025-09-07T07:36:03.8802064Z * [new branch] gh/ezyang/3127/head -> origin/gh/ezyang/3127/head 2025-09-07T07:36:03.8803402Z * [new branch] gh/ezyang/3127/orig -> origin/gh/ezyang/3127/orig 2025-09-07T07:36:03.8805274Z * [new branch] gh/ezyang/3128/base -> origin/gh/ezyang/3128/base 2025-09-07T07:36:03.8806468Z * [new branch] gh/ezyang/3128/head -> origin/gh/ezyang/3128/head 2025-09-07T07:36:03.8807633Z * [new branch] gh/ezyang/3128/orig -> origin/gh/ezyang/3128/orig 2025-09-07T07:36:03.8809366Z * [new branch] gh/ezyang/3129/base -> origin/gh/ezyang/3129/base 2025-09-07T07:36:03.8810535Z * [new branch] gh/ezyang/3129/head -> origin/gh/ezyang/3129/head 2025-09-07T07:36:03.8811793Z * [new branch] gh/ezyang/3129/orig -> origin/gh/ezyang/3129/orig 2025-09-07T07:36:03.8813700Z * [new branch] gh/ezyang/3130/base -> origin/gh/ezyang/3130/base 2025-09-07T07:36:03.8814902Z * [new branch] gh/ezyang/3130/head -> origin/gh/ezyang/3130/head 2025-09-07T07:36:03.8816096Z * [new branch] gh/ezyang/3130/orig -> origin/gh/ezyang/3130/orig 2025-09-07T07:36:03.8817869Z * [new branch] gh/ezyang/3131/base -> origin/gh/ezyang/3131/base 2025-09-07T07:36:03.8819133Z * [new branch] gh/ezyang/3131/head -> origin/gh/ezyang/3131/head 2025-09-07T07:36:03.8820366Z * [new branch] gh/ezyang/3131/orig -> origin/gh/ezyang/3131/orig 2025-09-07T07:36:03.8822143Z * [new branch] gh/ezyang/3132/base -> origin/gh/ezyang/3132/base 2025-09-07T07:36:03.8823299Z * [new branch] gh/ezyang/3132/head -> origin/gh/ezyang/3132/head 2025-09-07T07:36:03.8824476Z * [new branch] gh/ezyang/3132/orig -> origin/gh/ezyang/3132/orig 2025-09-07T07:36:03.8826180Z * [new branch] gh/ezyang/3133/base -> origin/gh/ezyang/3133/base 2025-09-07T07:36:03.8827362Z * [new branch] gh/ezyang/3133/head -> origin/gh/ezyang/3133/head 2025-09-07T07:36:03.8828546Z * [new branch] gh/ezyang/3133/orig -> origin/gh/ezyang/3133/orig 2025-09-07T07:36:03.8830383Z * [new branch] gh/ezyang/3134/base -> origin/gh/ezyang/3134/base 2025-09-07T07:36:03.8831544Z * [new branch] gh/ezyang/3134/head -> origin/gh/ezyang/3134/head 2025-09-07T07:36:03.8832691Z * [new branch] gh/ezyang/3134/orig -> origin/gh/ezyang/3134/orig 2025-09-07T07:36:03.8834495Z * [new branch] gh/ezyang/3135/base -> origin/gh/ezyang/3135/base 2025-09-07T07:36:03.8835673Z * [new branch] gh/ezyang/3135/head -> origin/gh/ezyang/3135/head 2025-09-07T07:36:03.8836960Z * [new branch] gh/ezyang/3135/orig -> origin/gh/ezyang/3135/orig 2025-09-07T07:36:03.8838663Z * [new branch] gh/ezyang/3136/base -> origin/gh/ezyang/3136/base 2025-09-07T07:36:03.8839917Z * [new branch] gh/ezyang/3136/head -> origin/gh/ezyang/3136/head 2025-09-07T07:36:03.8841018Z * [new branch] gh/ezyang/3136/orig -> origin/gh/ezyang/3136/orig 2025-09-07T07:36:03.8842769Z * [new branch] gh/ezyang/3137/base -> origin/gh/ezyang/3137/base 2025-09-07T07:36:03.8844102Z * [new branch] gh/ezyang/3137/head -> origin/gh/ezyang/3137/head 2025-09-07T07:36:03.8845259Z * [new branch] gh/ezyang/3137/orig -> origin/gh/ezyang/3137/orig 2025-09-07T07:36:03.8847126Z * [new branch] gh/ezyang/3138/base -> origin/gh/ezyang/3138/base 2025-09-07T07:36:03.8848600Z * [new branch] gh/ezyang/3138/head -> origin/gh/ezyang/3138/head 2025-09-07T07:36:03.8849961Z * [new branch] gh/ezyang/3138/orig -> origin/gh/ezyang/3138/orig 2025-09-07T07:36:03.8851807Z * [new branch] gh/ezyang/3139/base -> origin/gh/ezyang/3139/base 2025-09-07T07:36:03.8852977Z * [new branch] gh/ezyang/3139/head -> origin/gh/ezyang/3139/head 2025-09-07T07:36:03.8854200Z * [new branch] gh/ezyang/3139/orig -> origin/gh/ezyang/3139/orig 2025-09-07T07:36:03.8855957Z * [new branch] gh/ezyang/3140/base -> origin/gh/ezyang/3140/base 2025-09-07T07:36:03.8857127Z * [new branch] gh/ezyang/3140/head -> origin/gh/ezyang/3140/head 2025-09-07T07:36:03.8858478Z * [new branch] gh/ezyang/3140/orig -> origin/gh/ezyang/3140/orig 2025-09-07T07:36:03.8860223Z * [new branch] gh/ezyang/3141/base -> origin/gh/ezyang/3141/base 2025-09-07T07:36:03.8861423Z * [new branch] gh/ezyang/3141/head -> origin/gh/ezyang/3141/head 2025-09-07T07:36:03.8862624Z * [new branch] gh/ezyang/3141/orig -> origin/gh/ezyang/3141/orig 2025-09-07T07:36:03.8864434Z * [new branch] gh/ezyang/3142/base -> origin/gh/ezyang/3142/base 2025-09-07T07:36:03.8865598Z * [new branch] gh/ezyang/3142/head -> origin/gh/ezyang/3142/head 2025-09-07T07:36:03.8866778Z * [new branch] gh/ezyang/3142/orig -> origin/gh/ezyang/3142/orig 2025-09-07T07:36:03.8868486Z * [new branch] gh/ezyang/3143/base -> origin/gh/ezyang/3143/base 2025-09-07T07:36:03.8869679Z * [new branch] gh/ezyang/3143/head -> origin/gh/ezyang/3143/head 2025-09-07T07:36:03.8870859Z * [new branch] gh/ezyang/3143/orig -> origin/gh/ezyang/3143/orig 2025-09-07T07:36:03.8872975Z * [new branch] gh/fadara01/1/base -> origin/gh/fadara01/1/base 2025-09-07T07:36:03.8875177Z * [new branch] gh/fadara01/1/head -> origin/gh/fadara01/1/head 2025-09-07T07:36:03.8876385Z * [new branch] gh/fadara01/1/orig -> origin/gh/fadara01/1/orig 2025-09-07T07:36:03.8878652Z * [new branch] gh/fduwjj/171/base -> origin/gh/fduwjj/171/base 2025-09-07T07:36:03.8879932Z * [new branch] gh/fduwjj/171/head -> origin/gh/fduwjj/171/head 2025-09-07T07:36:03.8881181Z * [new branch] gh/fduwjj/171/orig -> origin/gh/fduwjj/171/orig 2025-09-07T07:36:03.8882999Z * [new branch] gh/fduwjj/175/base -> origin/gh/fduwjj/175/base 2025-09-07T07:36:03.8884325Z * [new branch] gh/fduwjj/175/head -> origin/gh/fduwjj/175/head 2025-09-07T07:36:03.8885638Z * [new branch] gh/fduwjj/175/orig -> origin/gh/fduwjj/175/orig 2025-09-07T07:36:03.8887461Z * [new branch] gh/fduwjj/176/base -> origin/gh/fduwjj/176/base 2025-09-07T07:36:03.8888635Z * [new branch] gh/fduwjj/176/head -> origin/gh/fduwjj/176/head 2025-09-07T07:36:03.8889847Z * [new branch] gh/fduwjj/176/orig -> origin/gh/fduwjj/176/orig 2025-09-07T07:36:03.8891608Z * [new branch] gh/fduwjj/177/base -> origin/gh/fduwjj/177/base 2025-09-07T07:36:03.8892900Z * [new branch] gh/fduwjj/177/head -> origin/gh/fduwjj/177/head 2025-09-07T07:36:03.8894019Z * [new branch] gh/fduwjj/177/orig -> origin/gh/fduwjj/177/orig 2025-09-07T07:36:03.8895703Z * [new branch] gh/fduwjj/178/base -> origin/gh/fduwjj/178/base 2025-09-07T07:36:03.8896976Z * [new branch] gh/fduwjj/178/head -> origin/gh/fduwjj/178/head 2025-09-07T07:36:03.8898330Z * [new branch] gh/fduwjj/178/orig -> origin/gh/fduwjj/178/orig 2025-09-07T07:36:03.8900044Z * [new branch] gh/fduwjj/179/base -> origin/gh/fduwjj/179/base 2025-09-07T07:36:03.8901188Z * [new branch] gh/fduwjj/179/head -> origin/gh/fduwjj/179/head 2025-09-07T07:36:03.8902394Z * [new branch] gh/fduwjj/179/orig -> origin/gh/fduwjj/179/orig 2025-09-07T07:36:03.8904329Z * [new branch] gh/fduwjj/180/base -> origin/gh/fduwjj/180/base 2025-09-07T07:36:03.8905537Z * [new branch] gh/fduwjj/180/head -> origin/gh/fduwjj/180/head 2025-09-07T07:36:03.8906722Z * [new branch] gh/fduwjj/180/orig -> origin/gh/fduwjj/180/orig 2025-09-07T07:36:03.8908428Z * [new branch] gh/fduwjj/181/base -> origin/gh/fduwjj/181/base 2025-09-07T07:36:03.8909598Z * [new branch] gh/fduwjj/181/head -> origin/gh/fduwjj/181/head 2025-09-07T07:36:03.8910787Z * [new branch] gh/fduwjj/181/orig -> origin/gh/fduwjj/181/orig 2025-09-07T07:36:03.8912473Z * [new branch] gh/fduwjj/182/base -> origin/gh/fduwjj/182/base 2025-09-07T07:36:03.8913655Z * [new branch] gh/fduwjj/182/head -> origin/gh/fduwjj/182/head 2025-09-07T07:36:03.8914913Z * [new branch] gh/fduwjj/182/orig -> origin/gh/fduwjj/182/orig 2025-09-07T07:36:03.8916680Z * [new branch] gh/fduwjj/183/base -> origin/gh/fduwjj/183/base 2025-09-07T07:36:03.8918021Z * [new branch] gh/fduwjj/183/head -> origin/gh/fduwjj/183/head 2025-09-07T07:36:03.8919220Z * [new branch] gh/fduwjj/183/orig -> origin/gh/fduwjj/183/orig 2025-09-07T07:36:03.8921293Z * [new branch] gh/fduwjj/184/base -> origin/gh/fduwjj/184/base 2025-09-07T07:36:03.8922468Z * [new branch] gh/fduwjj/184/head -> origin/gh/fduwjj/184/head 2025-09-07T07:36:03.8923654Z * [new branch] gh/fduwjj/184/orig -> origin/gh/fduwjj/184/orig 2025-09-07T07:36:03.8925436Z * [new branch] gh/fduwjj/185/base -> origin/gh/fduwjj/185/base 2025-09-07T07:36:03.8926603Z * [new branch] gh/fduwjj/185/head -> origin/gh/fduwjj/185/head 2025-09-07T07:36:03.8927812Z * [new branch] gh/fduwjj/185/orig -> origin/gh/fduwjj/185/orig 2025-09-07T07:36:03.8929406Z * [new branch] gh/fduwjj/186/base -> origin/gh/fduwjj/186/base 2025-09-07T07:36:03.8930610Z * [new branch] gh/fduwjj/186/head -> origin/gh/fduwjj/186/head 2025-09-07T07:36:03.8931796Z * [new branch] gh/fduwjj/186/orig -> origin/gh/fduwjj/186/orig 2025-09-07T07:36:03.8933471Z * [new branch] gh/fduwjj/187/base -> origin/gh/fduwjj/187/base 2025-09-07T07:36:03.8934742Z * [new branch] gh/fduwjj/187/head -> origin/gh/fduwjj/187/head 2025-09-07T07:36:03.8935931Z * [new branch] gh/fduwjj/187/orig -> origin/gh/fduwjj/187/orig 2025-09-07T07:36:03.8937569Z * [new branch] gh/fduwjj/188/base -> origin/gh/fduwjj/188/base 2025-09-07T07:36:03.8938883Z * [new branch] gh/fduwjj/188/head -> origin/gh/fduwjj/188/head 2025-09-07T07:36:03.8940031Z * [new branch] gh/fduwjj/188/orig -> origin/gh/fduwjj/188/orig 2025-09-07T07:36:03.8941750Z * [new branch] gh/fduwjj/189/base -> origin/gh/fduwjj/189/base 2025-09-07T07:36:03.8942759Z * [new branch] gh/fduwjj/189/head -> origin/gh/fduwjj/189/head 2025-09-07T07:36:03.8944388Z * [new branch] gh/fduwjj/189/orig -> origin/gh/fduwjj/189/orig 2025-09-07T07:36:03.8946513Z * [new branch] gh/fduwjj/190/base -> origin/gh/fduwjj/190/base 2025-09-07T07:36:03.8947724Z * [new branch] gh/fduwjj/190/head -> origin/gh/fduwjj/190/head 2025-09-07T07:36:03.8949031Z * [new branch] gh/fduwjj/190/orig -> origin/gh/fduwjj/190/orig 2025-09-07T07:36:03.8950646Z * [new branch] gh/fduwjj/191/base -> origin/gh/fduwjj/191/base 2025-09-07T07:36:03.8952042Z * [new branch] gh/fduwjj/191/head -> origin/gh/fduwjj/191/head 2025-09-07T07:36:03.8953481Z * [new branch] gh/fduwjj/191/orig -> origin/gh/fduwjj/191/orig 2025-09-07T07:36:03.8955750Z * [new branch] gh/fegin/306/base -> origin/gh/fegin/306/base 2025-09-07T07:36:03.8957582Z * [new branch] gh/fegin/306/head -> origin/gh/fegin/306/head 2025-09-07T07:36:03.8958815Z * [new branch] gh/fegin/306/orig -> origin/gh/fegin/306/orig 2025-09-07T07:36:03.8960492Z * [new branch] gh/fegin/307/base -> origin/gh/fegin/307/base 2025-09-07T07:36:03.8961692Z * [new branch] gh/fegin/307/head -> origin/gh/fegin/307/head 2025-09-07T07:36:03.8962894Z * [new branch] gh/fegin/307/orig -> origin/gh/fegin/307/orig 2025-09-07T07:36:03.8964650Z * [new branch] gh/fegin/308/base -> origin/gh/fegin/308/base 2025-09-07T07:36:03.8966725Z * [new branch] gh/fegin/308/head -> origin/gh/fegin/308/head 2025-09-07T07:36:03.8967176Z * [new branch] gh/fegin/308/orig -> origin/gh/fegin/308/orig 2025-09-07T07:36:03.8968858Z * [new branch] gh/fegin/309/base -> origin/gh/fegin/309/base 2025-09-07T07:36:03.8970061Z * [new branch] gh/fegin/309/head -> origin/gh/fegin/309/head 2025-09-07T07:36:03.8971310Z * [new branch] gh/fegin/309/orig -> origin/gh/fegin/309/orig 2025-09-07T07:36:03.8973191Z * [new branch] gh/fegin/310/base -> origin/gh/fegin/310/base 2025-09-07T07:36:03.8974220Z * [new branch] gh/fegin/310/head -> origin/gh/fegin/310/head 2025-09-07T07:36:03.8975455Z * [new branch] gh/fegin/310/orig -> origin/gh/fegin/310/orig 2025-09-07T07:36:03.8977220Z * [new branch] gh/fegin/311/base -> origin/gh/fegin/311/base 2025-09-07T07:36:03.8978518Z * [new branch] gh/fegin/311/head -> origin/gh/fegin/311/head 2025-09-07T07:36:03.8979804Z * [new branch] gh/fegin/311/orig -> origin/gh/fegin/311/orig 2025-09-07T07:36:03.8981439Z * [new branch] gh/fegin/312/base -> origin/gh/fegin/312/base 2025-09-07T07:36:03.8982617Z * [new branch] gh/fegin/312/head -> origin/gh/fegin/312/head 2025-09-07T07:36:03.8983905Z * [new branch] gh/fegin/312/orig -> origin/gh/fegin/312/orig 2025-09-07T07:36:03.8985588Z * [new branch] gh/fegin/313/base -> origin/gh/fegin/313/base 2025-09-07T07:36:03.8986782Z * [new branch] gh/fegin/313/head -> origin/gh/fegin/313/head 2025-09-07T07:36:03.8987971Z * [new branch] gh/fegin/313/orig -> origin/gh/fegin/313/orig 2025-09-07T07:36:03.8990125Z * [new branch] gh/fffrog/124/base -> origin/gh/fffrog/124/base 2025-09-07T07:36:03.8991398Z * [new branch] gh/fffrog/124/head -> origin/gh/fffrog/124/head 2025-09-07T07:36:03.8994796Z * [new branch] gh/fffrog/124/orig -> origin/gh/fffrog/124/orig 2025-09-07T07:36:03.8995762Z * [new branch] gh/fffrog/129/base -> origin/gh/fffrog/129/base 2025-09-07T07:36:03.8996554Z * [new branch] gh/fffrog/129/head -> origin/gh/fffrog/129/head 2025-09-07T07:36:03.8997458Z * [new branch] gh/fffrog/129/orig -> origin/gh/fffrog/129/orig 2025-09-07T07:36:03.8999186Z * [new branch] gh/fffrog/130/base -> origin/gh/fffrog/130/base 2025-09-07T07:36:03.9000447Z * [new branch] gh/fffrog/130/head -> origin/gh/fffrog/130/head 2025-09-07T07:36:03.9001711Z * [new branch] gh/fffrog/130/orig -> origin/gh/fffrog/130/orig 2025-09-07T07:36:03.9003608Z * [new branch] gh/fffrog/131/base -> origin/gh/fffrog/131/base 2025-09-07T07:36:03.9004855Z * [new branch] gh/fffrog/131/head -> origin/gh/fffrog/131/head 2025-09-07T07:36:03.9006074Z * [new branch] gh/fffrog/131/orig -> origin/gh/fffrog/131/orig 2025-09-07T07:36:03.9007742Z * [new branch] gh/fffrog/132/base -> origin/gh/fffrog/132/base 2025-09-07T07:36:03.9008927Z * [new branch] gh/fffrog/132/head -> origin/gh/fffrog/132/head 2025-09-07T07:36:03.9010100Z * [new branch] gh/fffrog/132/orig -> origin/gh/fffrog/132/orig 2025-09-07T07:36:03.9011829Z * [new branch] gh/fffrog/133/base -> origin/gh/fffrog/133/base 2025-09-07T07:36:03.9013007Z * [new branch] gh/fffrog/133/head -> origin/gh/fffrog/133/head 2025-09-07T07:36:03.9014221Z * [new branch] gh/fffrog/133/orig -> origin/gh/fffrog/133/orig 2025-09-07T07:36:03.9015858Z * [new branch] gh/fffrog/134/base -> origin/gh/fffrog/134/base 2025-09-07T07:36:03.9017103Z * [new branch] gh/fffrog/134/head -> origin/gh/fffrog/134/head 2025-09-07T07:36:03.9018838Z * [new branch] gh/fffrog/134/orig -> origin/gh/fffrog/134/orig 2025-09-07T07:36:03.9020670Z * [new branch] gh/fffrog/135/base -> origin/gh/fffrog/135/base 2025-09-07T07:36:03.9021895Z * [new branch] gh/fffrog/135/head -> origin/gh/fffrog/135/head 2025-09-07T07:36:03.9023127Z * [new branch] gh/fffrog/135/orig -> origin/gh/fffrog/135/orig 2025-09-07T07:36:03.9024970Z * [new branch] gh/fffrog/136/base -> origin/gh/fffrog/136/base 2025-09-07T07:36:03.9026009Z * [new branch] gh/fffrog/136/head -> origin/gh/fffrog/136/head 2025-09-07T07:36:03.9027194Z * [new branch] gh/fffrog/136/orig -> origin/gh/fffrog/136/orig 2025-09-07T07:36:03.9029007Z * [new branch] gh/fffrog/137/base -> origin/gh/fffrog/137/base 2025-09-07T07:36:03.9030140Z * [new branch] gh/fffrog/137/head -> origin/gh/fffrog/137/head 2025-09-07T07:36:03.9031347Z * [new branch] gh/fffrog/137/orig -> origin/gh/fffrog/137/orig 2025-09-07T07:36:03.9033009Z * [new branch] gh/fffrog/138/base -> origin/gh/fffrog/138/base 2025-09-07T07:36:03.9034250Z * [new branch] gh/fffrog/138/head -> origin/gh/fffrog/138/head 2025-09-07T07:36:03.9035451Z * [new branch] gh/fffrog/138/orig -> origin/gh/fffrog/138/orig 2025-09-07T07:36:03.9037137Z * [new branch] gh/fffrog/139/base -> origin/gh/fffrog/139/base 2025-09-07T07:36:03.9038356Z * [new branch] gh/fffrog/139/head -> origin/gh/fffrog/139/head 2025-09-07T07:36:03.9039581Z * [new branch] gh/fffrog/139/orig -> origin/gh/fffrog/139/orig 2025-09-07T07:36:03.9041316Z * [new branch] gh/fffrog/140/base -> origin/gh/fffrog/140/base 2025-09-07T07:36:03.9042479Z * [new branch] gh/fffrog/140/head -> origin/gh/fffrog/140/head 2025-09-07T07:36:03.9043611Z * [new branch] gh/fffrog/140/orig -> origin/gh/fffrog/140/orig 2025-09-07T07:36:03.9045375Z * [new branch] gh/fffrog/141/base -> origin/gh/fffrog/141/base 2025-09-07T07:36:03.9046471Z * [new branch] gh/fffrog/141/head -> origin/gh/fffrog/141/head 2025-09-07T07:36:03.9047610Z * [new branch] gh/fffrog/141/orig -> origin/gh/fffrog/141/orig 2025-09-07T07:36:03.9049294Z * [new branch] gh/fffrog/142/base -> origin/gh/fffrog/142/base 2025-09-07T07:36:03.9050611Z * [new branch] gh/fffrog/142/head -> origin/gh/fffrog/142/head 2025-09-07T07:36:03.9051789Z * [new branch] gh/fffrog/142/orig -> origin/gh/fffrog/142/orig 2025-09-07T07:36:03.9053437Z * [new branch] gh/fffrog/143/base -> origin/gh/fffrog/143/base 2025-09-07T07:36:03.9054641Z * [new branch] gh/fffrog/143/head -> origin/gh/fffrog/143/head 2025-09-07T07:36:03.9055837Z * [new branch] gh/fffrog/143/orig -> origin/gh/fffrog/143/orig 2025-09-07T07:36:03.9058053Z * [new branch] gh/fffrog/144/base -> origin/gh/fffrog/144/base 2025-09-07T07:36:03.9059242Z * [new branch] gh/fffrog/144/head -> origin/gh/fffrog/144/head 2025-09-07T07:36:03.9060419Z * [new branch] gh/fffrog/144/orig -> origin/gh/fffrog/144/orig 2025-09-07T07:36:03.9062710Z * [new branch] gh/fffrog/145/base -> origin/gh/fffrog/145/base 2025-09-07T07:36:03.9063403Z * [new branch] gh/fffrog/145/head -> origin/gh/fffrog/145/head 2025-09-07T07:36:03.9064762Z * [new branch] gh/fffrog/145/orig -> origin/gh/fffrog/145/orig 2025-09-07T07:36:03.9066488Z * [new branch] gh/fffrog/146/base -> origin/gh/fffrog/146/base 2025-09-07T07:36:03.9067800Z * [new branch] gh/fffrog/146/head -> origin/gh/fffrog/146/head 2025-09-07T07:36:03.9068973Z * [new branch] gh/fffrog/146/orig -> origin/gh/fffrog/146/orig 2025-09-07T07:36:03.9070761Z * [new branch] gh/fffrog/147/base -> origin/gh/fffrog/147/base 2025-09-07T07:36:03.9072138Z * [new branch] gh/fffrog/147/head -> origin/gh/fffrog/147/head 2025-09-07T07:36:03.9087050Z * [new branch] gh/fffrog/147/orig -> origin/gh/fffrog/147/orig 2025-09-07T07:36:03.9087260Z * [new branch] gh/fffrog/148/base -> origin/gh/fffrog/148/base 2025-09-07T07:36:03.9087410Z * [new branch] gh/fffrog/148/head -> origin/gh/fffrog/148/head 2025-09-07T07:36:03.9087542Z * [new branch] gh/fffrog/148/orig -> origin/gh/fffrog/148/orig 2025-09-07T07:36:03.9087662Z * [new branch] gh/fffrog/149/base -> origin/gh/fffrog/149/base 2025-09-07T07:36:03.9087791Z * [new branch] gh/fffrog/149/head -> origin/gh/fffrog/149/head 2025-09-07T07:36:03.9087911Z * [new branch] gh/fffrog/149/orig -> origin/gh/fffrog/149/orig 2025-09-07T07:36:03.9088030Z * [new branch] gh/fffrog/150/base -> origin/gh/fffrog/150/base 2025-09-07T07:36:03.9088153Z * [new branch] gh/fffrog/150/head -> origin/gh/fffrog/150/head 2025-09-07T07:36:03.9088270Z * [new branch] gh/fffrog/150/orig -> origin/gh/fffrog/150/orig 2025-09-07T07:36:03.9088396Z * [new branch] gh/fffrog/151/base -> origin/gh/fffrog/151/base 2025-09-07T07:36:03.9088519Z * [new branch] gh/fffrog/151/head -> origin/gh/fffrog/151/head 2025-09-07T07:36:03.9089656Z * [new branch] gh/fffrog/151/orig -> origin/gh/fffrog/151/orig 2025-09-07T07:36:03.9091469Z * [new branch] gh/fffrog/152/base -> origin/gh/fffrog/152/base 2025-09-07T07:36:03.9092704Z * [new branch] gh/fffrog/152/head -> origin/gh/fffrog/152/head 2025-09-07T07:36:03.9094548Z * [new branch] gh/fffrog/153/base -> origin/gh/fffrog/153/base 2025-09-07T07:36:03.9095644Z * [new branch] gh/fffrog/153/head -> origin/gh/fffrog/153/head 2025-09-07T07:36:03.9096827Z * [new branch] gh/fffrog/153/orig -> origin/gh/fffrog/153/orig 2025-09-07T07:36:03.9099493Z * [new branch] gh/gmagogsfm/1/base -> origin/gh/gmagogsfm/1/base 2025-09-07T07:36:03.9101002Z * [new branch] gh/gmagogsfm/1/head -> origin/gh/gmagogsfm/1/head 2025-09-07T07:36:03.9102108Z * [new branch] gh/gmagogsfm/1/orig -> origin/gh/gmagogsfm/1/orig 2025-09-07T07:36:03.9103953Z * [new branch] gh/gmagogsfm/2/base -> origin/gh/gmagogsfm/2/base 2025-09-07T07:36:03.9105202Z * [new branch] gh/gmagogsfm/2/head -> origin/gh/gmagogsfm/2/head 2025-09-07T07:36:03.9106384Z * [new branch] gh/gmagogsfm/2/orig -> origin/gh/gmagogsfm/2/orig 2025-09-07T07:36:03.9108002Z * [new branch] gh/gmagogsfm/3/base -> origin/gh/gmagogsfm/3/base 2025-09-07T07:36:03.9109264Z * [new branch] gh/gmagogsfm/3/head -> origin/gh/gmagogsfm/3/head 2025-09-07T07:36:03.9110443Z * [new branch] gh/gmagogsfm/3/orig -> origin/gh/gmagogsfm/3/orig 2025-09-07T07:36:03.9112596Z * [new branch] gh/guangyey/134/base -> origin/gh/guangyey/134/base 2025-09-07T07:36:03.9113772Z * [new branch] gh/guangyey/134/head -> origin/gh/guangyey/134/head 2025-09-07T07:36:03.9114956Z * [new branch] gh/guangyey/134/orig -> origin/gh/guangyey/134/orig 2025-09-07T07:36:03.9116718Z * [new branch] gh/guangyey/135/base -> origin/gh/guangyey/135/base 2025-09-07T07:36:03.9117910Z * [new branch] gh/guangyey/135/head -> origin/gh/guangyey/135/head 2025-09-07T07:36:03.9119087Z * [new branch] gh/guangyey/135/orig -> origin/gh/guangyey/135/orig 2025-09-07T07:36:03.9120786Z * [new branch] gh/guangyey/139/base -> origin/gh/guangyey/139/base 2025-09-07T07:36:03.9122056Z * [new branch] gh/guangyey/139/head -> origin/gh/guangyey/139/head 2025-09-07T07:36:03.9123215Z * [new branch] gh/guangyey/139/orig -> origin/gh/guangyey/139/orig 2025-09-07T07:36:03.9124910Z * [new branch] gh/guangyey/140/base -> origin/gh/guangyey/140/base 2025-09-07T07:36:03.9126083Z * [new branch] gh/guangyey/140/head -> origin/gh/guangyey/140/head 2025-09-07T07:36:03.9127256Z * [new branch] gh/guangyey/140/orig -> origin/gh/guangyey/140/orig 2025-09-07T07:36:03.9128924Z * [new branch] gh/guangyey/142/base -> origin/gh/guangyey/142/base 2025-09-07T07:36:03.9130087Z * [new branch] gh/guangyey/142/head -> origin/gh/guangyey/142/head 2025-09-07T07:36:03.9131289Z * [new branch] gh/guangyey/142/orig -> origin/gh/guangyey/142/orig 2025-09-07T07:36:03.9133025Z * [new branch] gh/guangyey/145/base -> origin/gh/guangyey/145/base 2025-09-07T07:36:03.9134220Z * [new branch] gh/guangyey/145/head -> origin/gh/guangyey/145/head 2025-09-07T07:36:03.9135512Z * [new branch] gh/guangyey/145/orig -> origin/gh/guangyey/145/orig 2025-09-07T07:36:03.9137245Z * [new branch] gh/guangyey/153/base -> origin/gh/guangyey/153/base 2025-09-07T07:36:03.9138471Z * [new branch] gh/guangyey/153/head -> origin/gh/guangyey/153/head 2025-09-07T07:36:03.9139652Z * [new branch] gh/guangyey/153/orig -> origin/gh/guangyey/153/orig 2025-09-07T07:36:03.9141348Z * [new branch] gh/guangyey/159/base -> origin/gh/guangyey/159/base 2025-09-07T07:36:03.9142541Z * [new branch] gh/guangyey/159/head -> origin/gh/guangyey/159/head 2025-09-07T07:36:03.9143825Z * [new branch] gh/guangyey/159/orig -> origin/gh/guangyey/159/orig 2025-09-07T07:36:03.9145521Z * [new branch] gh/guangyey/163/base -> origin/gh/guangyey/163/base 2025-09-07T07:36:03.9146701Z * [new branch] gh/guangyey/163/head -> origin/gh/guangyey/163/head 2025-09-07T07:36:03.9147862Z * [new branch] gh/guangyey/163/orig -> origin/gh/guangyey/163/orig 2025-09-07T07:36:03.9149661Z * [new branch] gh/guangyey/168/base -> origin/gh/guangyey/168/base 2025-09-07T07:36:03.9150842Z * [new branch] gh/guangyey/168/head -> origin/gh/guangyey/168/head 2025-09-07T07:36:03.9152065Z * [new branch] gh/guangyey/168/orig -> origin/gh/guangyey/168/orig 2025-09-07T07:36:03.9153724Z * [new branch] gh/guangyey/169/base -> origin/gh/guangyey/169/base 2025-09-07T07:36:03.9154920Z * [new branch] gh/guangyey/169/head -> origin/gh/guangyey/169/head 2025-09-07T07:36:03.9156082Z * [new branch] gh/guangyey/169/orig -> origin/gh/guangyey/169/orig 2025-09-07T07:36:03.9157792Z * [new branch] gh/guangyey/170/base -> origin/gh/guangyey/170/base 2025-09-07T07:36:03.9158958Z * [new branch] gh/guangyey/170/head -> origin/gh/guangyey/170/head 2025-09-07T07:36:03.9160171Z * [new branch] gh/guangyey/170/orig -> origin/gh/guangyey/170/orig 2025-09-07T07:36:03.9161843Z * [new branch] gh/guangyey/171/base -> origin/gh/guangyey/171/base 2025-09-07T07:36:03.9163050Z * [new branch] gh/guangyey/171/head -> origin/gh/guangyey/171/head 2025-09-07T07:36:03.9164229Z * [new branch] gh/guangyey/171/orig -> origin/gh/guangyey/171/orig 2025-09-07T07:36:03.9166404Z * [new branch] gh/guangyey/174/base -> origin/gh/guangyey/174/base 2025-09-07T07:36:03.9167607Z * [new branch] gh/guangyey/174/head -> origin/gh/guangyey/174/head 2025-09-07T07:36:03.9168776Z * [new branch] gh/guangyey/174/orig -> origin/gh/guangyey/174/orig 2025-09-07T07:36:03.9170598Z * [new branch] gh/guangyey/176/base -> origin/gh/guangyey/176/base 2025-09-07T07:36:03.9171766Z * [new branch] gh/guangyey/176/head -> origin/gh/guangyey/176/head 2025-09-07T07:36:03.9172968Z * [new branch] gh/guangyey/176/orig -> origin/gh/guangyey/176/orig 2025-09-07T07:36:03.9174636Z * [new branch] gh/guangyey/178/base -> origin/gh/guangyey/178/base 2025-09-07T07:36:03.9175820Z * [new branch] gh/guangyey/178/head -> origin/gh/guangyey/178/head 2025-09-07T07:36:03.9177002Z * [new branch] gh/guangyey/178/orig -> origin/gh/guangyey/178/orig 2025-09-07T07:36:03.9179214Z * [new branch] gh/guangyey/181/base -> origin/gh/guangyey/181/base 2025-09-07T07:36:03.9180432Z * [new branch] gh/guangyey/181/head -> origin/gh/guangyey/181/head 2025-09-07T07:36:03.9181597Z * [new branch] gh/guangyey/181/orig -> origin/gh/guangyey/181/orig 2025-09-07T07:36:03.9183393Z * [new branch] gh/guangyey/182/base -> origin/gh/guangyey/182/base 2025-09-07T07:36:03.9184560Z * [new branch] gh/guangyey/182/head -> origin/gh/guangyey/182/head 2025-09-07T07:36:03.9185744Z * [new branch] gh/guangyey/182/orig -> origin/gh/guangyey/182/orig 2025-09-07T07:36:03.9187490Z * [new branch] gh/guangyey/183/base -> origin/gh/guangyey/183/base 2025-09-07T07:36:03.9188699Z * [new branch] gh/guangyey/183/head -> origin/gh/guangyey/183/head 2025-09-07T07:36:03.9189940Z * [new branch] gh/guangyey/183/orig -> origin/gh/guangyey/183/orig 2025-09-07T07:36:03.9191741Z * [new branch] gh/guangyey/184/base -> origin/gh/guangyey/184/base 2025-09-07T07:36:03.9192998Z * [new branch] gh/guangyey/184/head -> origin/gh/guangyey/184/head 2025-09-07T07:36:03.9194141Z * [new branch] gh/guangyey/184/orig -> origin/gh/guangyey/184/orig 2025-09-07T07:36:03.9195845Z * [new branch] gh/guangyey/185/base -> origin/gh/guangyey/185/base 2025-09-07T07:36:03.9197034Z * [new branch] gh/guangyey/185/head -> origin/gh/guangyey/185/head 2025-09-07T07:36:03.9198245Z * [new branch] gh/guangyey/185/orig -> origin/gh/guangyey/185/orig 2025-09-07T07:36:03.9200057Z * [new branch] gh/guangyey/186/base -> origin/gh/guangyey/186/base 2025-09-07T07:36:03.9201327Z * [new branch] gh/guangyey/186/head -> origin/gh/guangyey/186/head 2025-09-07T07:36:03.9202421Z * [new branch] gh/guangyey/186/orig -> origin/gh/guangyey/186/orig 2025-09-07T07:36:03.9204399Z * [new branch] gh/guangyey/187/base -> origin/gh/guangyey/187/base 2025-09-07T07:36:03.9205635Z * [new branch] gh/guangyey/187/head -> origin/gh/guangyey/187/head 2025-09-07T07:36:03.9207022Z * [new branch] gh/guangyey/187/orig -> origin/gh/guangyey/187/orig 2025-09-07T07:36:03.9208771Z * [new branch] gh/guangyey/188/base -> origin/gh/guangyey/188/base 2025-09-07T07:36:03.9209992Z * [new branch] gh/guangyey/188/head -> origin/gh/guangyey/188/head 2025-09-07T07:36:03.9211171Z * [new branch] gh/guangyey/188/orig -> origin/gh/guangyey/188/orig 2025-09-07T07:36:03.9212928Z * [new branch] gh/guangyey/189/base -> origin/gh/guangyey/189/base 2025-09-07T07:36:03.9214155Z * [new branch] gh/guangyey/189/head -> origin/gh/guangyey/189/head 2025-09-07T07:36:03.9215347Z * [new branch] gh/guangyey/189/orig -> origin/gh/guangyey/189/orig 2025-09-07T07:36:03.9217134Z * [new branch] gh/guangyey/190/base -> origin/gh/guangyey/190/base 2025-09-07T07:36:03.9218447Z * [new branch] gh/guangyey/190/head -> origin/gh/guangyey/190/head 2025-09-07T07:36:03.9219668Z * [new branch] gh/guangyey/190/orig -> origin/gh/guangyey/190/orig 2025-09-07T07:36:03.9221404Z * [new branch] gh/guangyey/191/base -> origin/gh/guangyey/191/base 2025-09-07T07:36:03.9222639Z * [new branch] gh/guangyey/191/head -> origin/gh/guangyey/191/head 2025-09-07T07:36:03.9223810Z * [new branch] gh/guangyey/191/orig -> origin/gh/guangyey/191/orig 2025-09-07T07:36:03.9225575Z * [new branch] gh/guangyey/192/base -> origin/gh/guangyey/192/base 2025-09-07T07:36:03.9226752Z * [new branch] gh/guangyey/192/head -> origin/gh/guangyey/192/head 2025-09-07T07:36:03.9228021Z * [new branch] gh/guangyey/192/orig -> origin/gh/guangyey/192/orig 2025-09-07T07:36:03.9229835Z * [new branch] gh/guangyey/193/base -> origin/gh/guangyey/193/base 2025-09-07T07:36:03.9231015Z * [new branch] gh/guangyey/193/head -> origin/gh/guangyey/193/head 2025-09-07T07:36:03.9232195Z * [new branch] gh/guangyey/193/orig -> origin/gh/guangyey/193/orig 2025-09-07T07:36:03.9234461Z * [new branch] gh/guangyey/194/base -> origin/gh/guangyey/194/base 2025-09-07T07:36:03.9235662Z * [new branch] gh/guangyey/194/head -> origin/gh/guangyey/194/head 2025-09-07T07:36:03.9236851Z * [new branch] gh/guangyey/194/orig -> origin/gh/guangyey/194/orig 2025-09-07T07:36:03.9238587Z * [new branch] gh/guangyey/195/base -> origin/gh/guangyey/195/base 2025-09-07T07:36:03.9239841Z * [new branch] gh/guangyey/195/head -> origin/gh/guangyey/195/head 2025-09-07T07:36:03.9241012Z * [new branch] gh/guangyey/195/orig -> origin/gh/guangyey/195/orig 2025-09-07T07:36:03.9243021Z * [new branch] gh/guangyey/196/base -> origin/gh/guangyey/196/base 2025-09-07T07:36:03.9244137Z * [new branch] gh/guangyey/196/head -> origin/gh/guangyey/196/head 2025-09-07T07:36:03.9245280Z * [new branch] gh/guangyey/196/orig -> origin/gh/guangyey/196/orig 2025-09-07T07:36:03.9246975Z * [new branch] gh/guangyey/197/base -> origin/gh/guangyey/197/base 2025-09-07T07:36:03.9248179Z * [new branch] gh/guangyey/197/head -> origin/gh/guangyey/197/head 2025-09-07T07:36:03.9249394Z * [new branch] gh/guangyey/197/orig -> origin/gh/guangyey/197/orig 2025-09-07T07:36:03.9251223Z * [new branch] gh/guangyey/198/base -> origin/gh/guangyey/198/base 2025-09-07T07:36:03.9252651Z * [new branch] gh/guangyey/198/head -> origin/gh/guangyey/198/head 2025-09-07T07:36:03.9253644Z * [new branch] gh/guangyey/198/orig -> origin/gh/guangyey/198/orig 2025-09-07T07:36:03.9255404Z * [new branch] gh/guangyey/199/base -> origin/gh/guangyey/199/base 2025-09-07T07:36:03.9257146Z * [new branch] gh/guangyey/199/head -> origin/gh/guangyey/199/head 2025-09-07T07:36:03.9258469Z * [new branch] gh/guangyey/199/orig -> origin/gh/guangyey/199/orig 2025-09-07T07:36:03.9260235Z * [new branch] gh/guangyey/200/base -> origin/gh/guangyey/200/base 2025-09-07T07:36:03.9261383Z * [new branch] gh/guangyey/200/head -> origin/gh/guangyey/200/head 2025-09-07T07:36:03.9262733Z * [new branch] gh/guangyey/200/orig -> origin/gh/guangyey/200/orig 2025-09-07T07:36:03.9264274Z * [new branch] gh/guangyey/201/base -> origin/gh/guangyey/201/base 2025-09-07T07:36:03.9265475Z * [new branch] gh/guangyey/201/head -> origin/gh/guangyey/201/head 2025-09-07T07:36:03.9266662Z * [new branch] gh/guangyey/201/orig -> origin/gh/guangyey/201/orig 2025-09-07T07:36:03.9268414Z * [new branch] gh/guangyey/202/base -> origin/gh/guangyey/202/base 2025-09-07T07:36:03.9269755Z * [new branch] gh/guangyey/202/head -> origin/gh/guangyey/202/head 2025-09-07T07:36:03.9270921Z * [new branch] gh/guangyey/202/orig -> origin/gh/guangyey/202/orig 2025-09-07T07:36:03.9272651Z * [new branch] gh/guangyey/203/base -> origin/gh/guangyey/203/base 2025-09-07T07:36:03.9273840Z * [new branch] gh/guangyey/203/head -> origin/gh/guangyey/203/head 2025-09-07T07:36:03.9275013Z * [new branch] gh/guangyey/203/orig -> origin/gh/guangyey/203/orig 2025-09-07T07:36:03.9276735Z * [new branch] gh/guangyey/204/base -> origin/gh/guangyey/204/base 2025-09-07T07:36:03.9278035Z * [new branch] gh/guangyey/204/head -> origin/gh/guangyey/204/head 2025-09-07T07:36:03.9279217Z * [new branch] gh/guangyey/204/orig -> origin/gh/guangyey/204/orig 2025-09-07T07:36:03.9280917Z * [new branch] gh/guangyey/205/base -> origin/gh/guangyey/205/base 2025-09-07T07:36:03.9282105Z * [new branch] gh/guangyey/205/head -> origin/gh/guangyey/205/head 2025-09-07T07:36:03.9283251Z * [new branch] gh/guangyey/205/orig -> origin/gh/guangyey/205/orig 2025-09-07T07:36:03.9285049Z * [new branch] gh/guangyey/206/base -> origin/gh/guangyey/206/base 2025-09-07T07:36:03.9286294Z * [new branch] gh/guangyey/206/head -> origin/gh/guangyey/206/head 2025-09-07T07:36:03.9287471Z * [new branch] gh/guangyey/206/orig -> origin/gh/guangyey/206/orig 2025-09-07T07:36:03.9289222Z * [new branch] gh/guangyey/207/base -> origin/gh/guangyey/207/base 2025-09-07T07:36:03.9290427Z * [new branch] gh/guangyey/207/head -> origin/gh/guangyey/207/head 2025-09-07T07:36:03.9291661Z * [new branch] gh/guangyey/207/orig -> origin/gh/guangyey/207/orig 2025-09-07T07:36:03.9293302Z * [new branch] gh/guangyey/79/base -> origin/gh/guangyey/79/base 2025-09-07T07:36:03.9294494Z * [new branch] gh/guangyey/79/head -> origin/gh/guangyey/79/head 2025-09-07T07:36:03.9295650Z * [new branch] gh/guangyey/79/orig -> origin/gh/guangyey/79/orig 2025-09-07T07:36:03.9297428Z * [new branch] gh/guangyey/89/base -> origin/gh/guangyey/89/base 2025-09-07T07:36:03.9298712Z * [new branch] gh/guangyey/89/head -> origin/gh/guangyey/89/head 2025-09-07T07:36:03.9299977Z * [new branch] gh/guangyey/89/orig -> origin/gh/guangyey/89/orig 2025-09-07T07:36:03.9302289Z * [new branch] gh/guilhermeleobas/107/base -> origin/gh/guilhermeleobas/107/base 2025-09-07T07:36:03.9303675Z * [new branch] gh/guilhermeleobas/107/head -> origin/gh/guilhermeleobas/107/head 2025-09-07T07:36:03.9304927Z * [new branch] gh/guilhermeleobas/107/orig -> origin/gh/guilhermeleobas/107/orig 2025-09-07T07:36:03.9306752Z * [new branch] gh/guilhermeleobas/108/base -> origin/gh/guilhermeleobas/108/base 2025-09-07T07:36:03.9307777Z * [new branch] gh/guilhermeleobas/108/head -> origin/gh/guilhermeleobas/108/head 2025-09-07T07:36:03.9308948Z * [new branch] gh/guilhermeleobas/108/orig -> origin/gh/guilhermeleobas/108/orig 2025-09-07T07:36:03.9310685Z * [new branch] gh/guilhermeleobas/124/base -> origin/gh/guilhermeleobas/124/base 2025-09-07T07:36:03.9312036Z * [new branch] gh/guilhermeleobas/124/head -> origin/gh/guilhermeleobas/124/head 2025-09-07T07:36:03.9313419Z * [new branch] gh/guilhermeleobas/124/orig -> origin/gh/guilhermeleobas/124/orig 2025-09-07T07:36:03.9315128Z * [new branch] gh/guilhermeleobas/147/base -> origin/gh/guilhermeleobas/147/base 2025-09-07T07:36:03.9316307Z * [new branch] gh/guilhermeleobas/147/head -> origin/gh/guilhermeleobas/147/head 2025-09-07T07:36:03.9317509Z * [new branch] gh/guilhermeleobas/147/orig -> origin/gh/guilhermeleobas/147/orig 2025-09-07T07:36:03.9319348Z * [new branch] gh/guilhermeleobas/150/base -> origin/gh/guilhermeleobas/150/base 2025-09-07T07:36:03.9320956Z * [new branch] gh/guilhermeleobas/150/head -> origin/gh/guilhermeleobas/150/head 2025-09-07T07:36:03.9322166Z * [new branch] gh/guilhermeleobas/150/orig -> origin/gh/guilhermeleobas/150/orig 2025-09-07T07:36:03.9323909Z * [new branch] gh/guilhermeleobas/163/base -> origin/gh/guilhermeleobas/163/base 2025-09-07T07:36:03.9325139Z * [new branch] gh/guilhermeleobas/163/head -> origin/gh/guilhermeleobas/163/head 2025-09-07T07:36:03.9326326Z * [new branch] gh/guilhermeleobas/163/orig -> origin/gh/guilhermeleobas/163/orig 2025-09-07T07:36:03.9328345Z * [new branch] gh/guilhermeleobas/164/base -> origin/gh/guilhermeleobas/164/base 2025-09-07T07:36:03.9329236Z * [new branch] gh/guilhermeleobas/164/head -> origin/gh/guilhermeleobas/164/head 2025-09-07T07:36:03.9330435Z * [new branch] gh/guilhermeleobas/164/orig -> origin/gh/guilhermeleobas/164/orig 2025-09-07T07:36:03.9332112Z * [new branch] gh/guilhermeleobas/165/base -> origin/gh/guilhermeleobas/165/base 2025-09-07T07:36:03.9333321Z * [new branch] gh/guilhermeleobas/165/head -> origin/gh/guilhermeleobas/165/head 2025-09-07T07:36:03.9334540Z * [new branch] gh/guilhermeleobas/165/orig -> origin/gh/guilhermeleobas/165/orig 2025-09-07T07:36:03.9336310Z * [new branch] gh/guilhermeleobas/166/base -> origin/gh/guilhermeleobas/166/base 2025-09-07T07:36:03.9337513Z * [new branch] gh/guilhermeleobas/166/head -> origin/gh/guilhermeleobas/166/head 2025-09-07T07:36:03.9339229Z * [new branch] gh/guilhermeleobas/166/orig -> origin/gh/guilhermeleobas/166/orig 2025-09-07T07:36:03.9341096Z * [new branch] gh/guilhermeleobas/167/base -> origin/gh/guilhermeleobas/167/base 2025-09-07T07:36:03.9342186Z * [new branch] gh/guilhermeleobas/167/head -> origin/gh/guilhermeleobas/167/head 2025-09-07T07:36:03.9343400Z * [new branch] gh/guilhermeleobas/167/orig -> origin/gh/guilhermeleobas/167/orig 2025-09-07T07:36:03.9345085Z * [new branch] gh/guilhermeleobas/168/base -> origin/gh/guilhermeleobas/168/base 2025-09-07T07:36:03.9346273Z * [new branch] gh/guilhermeleobas/168/head -> origin/gh/guilhermeleobas/168/head 2025-09-07T07:36:03.9347451Z * [new branch] gh/guilhermeleobas/168/orig -> origin/gh/guilhermeleobas/168/orig 2025-09-07T07:36:03.9349274Z * [new branch] gh/guilhermeleobas/169/base -> origin/gh/guilhermeleobas/169/base 2025-09-07T07:36:03.9350490Z * [new branch] gh/guilhermeleobas/169/head -> origin/gh/guilhermeleobas/169/head 2025-09-07T07:36:03.9351715Z * [new branch] gh/guilhermeleobas/169/orig -> origin/gh/guilhermeleobas/169/orig 2025-09-07T07:36:03.9353481Z * [new branch] gh/guilhermeleobas/170/base -> origin/gh/guilhermeleobas/170/base 2025-09-07T07:36:03.9354763Z * [new branch] gh/guilhermeleobas/170/head -> origin/gh/guilhermeleobas/170/head 2025-09-07T07:36:03.9355975Z * [new branch] gh/guilhermeleobas/170/orig -> origin/gh/guilhermeleobas/170/orig 2025-09-07T07:36:03.9357668Z * [new branch] gh/guilhermeleobas/171/base -> origin/gh/guilhermeleobas/171/base 2025-09-07T07:36:03.9358870Z * [new branch] gh/guilhermeleobas/171/head -> origin/gh/guilhermeleobas/171/head 2025-09-07T07:36:03.9360054Z * [new branch] gh/guilhermeleobas/171/orig -> origin/gh/guilhermeleobas/171/orig 2025-09-07T07:36:03.9361736Z * [new branch] gh/guilhermeleobas/173/base -> origin/gh/guilhermeleobas/173/base 2025-09-07T07:36:03.9362894Z * [new branch] gh/guilhermeleobas/173/head -> origin/gh/guilhermeleobas/173/head 2025-09-07T07:36:03.9364081Z * [new branch] gh/guilhermeleobas/173/orig -> origin/gh/guilhermeleobas/173/orig 2025-09-07T07:36:03.9365753Z * [new branch] gh/guilhermeleobas/192/base -> origin/gh/guilhermeleobas/192/base 2025-09-07T07:36:03.9367016Z * [new branch] gh/guilhermeleobas/192/head -> origin/gh/guilhermeleobas/192/head 2025-09-07T07:36:03.9368192Z * [new branch] gh/guilhermeleobas/192/orig -> origin/gh/guilhermeleobas/192/orig 2025-09-07T07:36:03.9370324Z * [new branch] gh/guilhermeleobas/193/base -> origin/gh/guilhermeleobas/193/base 2025-09-07T07:36:03.9371556Z * [new branch] gh/guilhermeleobas/193/head -> origin/gh/guilhermeleobas/193/head 2025-09-07T07:36:03.9372751Z * [new branch] gh/guilhermeleobas/193/orig -> origin/gh/guilhermeleobas/193/orig 2025-09-07T07:36:03.9374493Z * [new branch] gh/guilhermeleobas/194/base -> origin/gh/guilhermeleobas/194/base 2025-09-07T07:36:03.9375673Z * [new branch] gh/guilhermeleobas/194/head -> origin/gh/guilhermeleobas/194/head 2025-09-07T07:36:03.9376907Z * [new branch] gh/guilhermeleobas/194/orig -> origin/gh/guilhermeleobas/194/orig 2025-09-07T07:36:03.9378751Z * [new branch] gh/guilhermeleobas/203/base -> origin/gh/guilhermeleobas/203/base 2025-09-07T07:36:03.9379967Z * [new branch] gh/guilhermeleobas/203/head -> origin/gh/guilhermeleobas/203/head 2025-09-07T07:36:03.9381121Z * [new branch] gh/guilhermeleobas/203/orig -> origin/gh/guilhermeleobas/203/orig 2025-09-07T07:36:03.9382840Z * [new branch] gh/guilhermeleobas/204/base -> origin/gh/guilhermeleobas/204/base 2025-09-07T07:36:03.9384190Z * [new branch] gh/guilhermeleobas/204/head -> origin/gh/guilhermeleobas/204/head 2025-09-07T07:36:03.9385357Z * [new branch] gh/guilhermeleobas/204/orig -> origin/gh/guilhermeleobas/204/orig 2025-09-07T07:36:03.9387676Z * [new branch] gh/guilhermeleobas/205/base -> origin/gh/guilhermeleobas/205/base 2025-09-07T07:36:03.9388866Z * [new branch] gh/guilhermeleobas/205/head -> origin/gh/guilhermeleobas/205/head 2025-09-07T07:36:03.9390091Z * [new branch] gh/guilhermeleobas/205/orig -> origin/gh/guilhermeleobas/205/orig 2025-09-07T07:36:03.9392054Z * [new branch] gh/guilhermeleobas/209/base -> origin/gh/guilhermeleobas/209/base 2025-09-07T07:36:03.9393250Z * [new branch] gh/guilhermeleobas/209/head -> origin/gh/guilhermeleobas/209/head 2025-09-07T07:36:03.9394440Z * [new branch] gh/guilhermeleobas/209/orig -> origin/gh/guilhermeleobas/209/orig 2025-09-07T07:36:03.9396289Z * [new branch] gh/guilhermeleobas/210/base -> origin/gh/guilhermeleobas/210/base 2025-09-07T07:36:03.9397470Z * [new branch] gh/guilhermeleobas/210/head -> origin/gh/guilhermeleobas/210/head 2025-09-07T07:36:03.9398654Z * [new branch] gh/guilhermeleobas/210/orig -> origin/gh/guilhermeleobas/210/orig 2025-09-07T07:36:03.9400418Z * [new branch] gh/guilhermeleobas/211/base -> origin/gh/guilhermeleobas/211/base 2025-09-07T07:36:03.9401651Z * [new branch] gh/guilhermeleobas/211/head -> origin/gh/guilhermeleobas/211/head 2025-09-07T07:36:03.9402830Z * [new branch] gh/guilhermeleobas/211/orig -> origin/gh/guilhermeleobas/211/orig 2025-09-07T07:36:03.9405014Z * [new branch] gh/guilhermeleobas/214/base -> origin/gh/guilhermeleobas/214/base 2025-09-07T07:36:03.9406234Z * [new branch] gh/guilhermeleobas/214/head -> origin/gh/guilhermeleobas/214/head 2025-09-07T07:36:03.9407468Z * [new branch] gh/guilhermeleobas/214/orig -> origin/gh/guilhermeleobas/214/orig 2025-09-07T07:36:03.9409193Z * [new branch] gh/guilhermeleobas/215/base -> origin/gh/guilhermeleobas/215/base 2025-09-07T07:36:03.9410401Z * [new branch] gh/guilhermeleobas/215/head -> origin/gh/guilhermeleobas/215/head 2025-09-07T07:36:03.9411609Z * [new branch] gh/guilhermeleobas/215/orig -> origin/gh/guilhermeleobas/215/orig 2025-09-07T07:36:03.9413319Z * [new branch] gh/guilhermeleobas/216/base -> origin/gh/guilhermeleobas/216/base 2025-09-07T07:36:03.9414523Z * [new branch] gh/guilhermeleobas/216/head -> origin/gh/guilhermeleobas/216/head 2025-09-07T07:36:03.9415704Z * [new branch] gh/guilhermeleobas/216/orig -> origin/gh/guilhermeleobas/216/orig 2025-09-07T07:36:03.9417516Z * [new branch] gh/guilhermeleobas/217/base -> origin/gh/guilhermeleobas/217/base 2025-09-07T07:36:03.9418845Z * [new branch] gh/guilhermeleobas/217/head -> origin/gh/guilhermeleobas/217/head 2025-09-07T07:36:03.9420212Z * [new branch] gh/guilhermeleobas/217/orig -> origin/gh/guilhermeleobas/217/orig 2025-09-07T07:36:03.9422045Z * [new branch] gh/guilhermeleobas/219/base -> origin/gh/guilhermeleobas/219/base 2025-09-07T07:36:03.9423240Z * [new branch] gh/guilhermeleobas/219/head -> origin/gh/guilhermeleobas/219/head 2025-09-07T07:36:03.9424434Z * [new branch] gh/guilhermeleobas/219/orig -> origin/gh/guilhermeleobas/219/orig 2025-09-07T07:36:03.9426150Z * [new branch] gh/guilhermeleobas/220/base -> origin/gh/guilhermeleobas/220/base 2025-09-07T07:36:03.9427337Z * [new branch] gh/guilhermeleobas/220/head -> origin/gh/guilhermeleobas/220/head 2025-09-07T07:36:03.9428497Z * [new branch] gh/guilhermeleobas/220/orig -> origin/gh/guilhermeleobas/220/orig 2025-09-07T07:36:03.9430257Z * [new branch] gh/guilhermeleobas/221/base -> origin/gh/guilhermeleobas/221/base 2025-09-07T07:36:03.9431438Z * [new branch] gh/guilhermeleobas/221/head -> origin/gh/guilhermeleobas/221/head 2025-09-07T07:36:03.9432770Z * [new branch] gh/guilhermeleobas/221/orig -> origin/gh/guilhermeleobas/221/orig 2025-09-07T07:36:03.9434627Z * [new branch] gh/guilhermeleobas/222/base -> origin/gh/guilhermeleobas/222/base 2025-09-07T07:36:03.9435626Z * [new branch] gh/guilhermeleobas/222/head -> origin/gh/guilhermeleobas/222/head 2025-09-07T07:36:03.9436780Z * [new branch] gh/guilhermeleobas/222/orig -> origin/gh/guilhermeleobas/222/orig 2025-09-07T07:36:03.9438692Z * [new branch] gh/guilhermeleobas/223/base -> origin/gh/guilhermeleobas/223/base 2025-09-07T07:36:03.9439919Z * [new branch] gh/guilhermeleobas/223/head -> origin/gh/guilhermeleobas/223/head 2025-09-07T07:36:03.9441195Z * [new branch] gh/guilhermeleobas/223/orig -> origin/gh/guilhermeleobas/223/orig 2025-09-07T07:36:03.9442941Z * [new branch] gh/guilhermeleobas/224/base -> origin/gh/guilhermeleobas/224/base 2025-09-07T07:36:03.9444162Z * [new branch] gh/guilhermeleobas/224/head -> origin/gh/guilhermeleobas/224/head 2025-09-07T07:36:03.9445334Z * [new branch] gh/guilhermeleobas/224/orig -> origin/gh/guilhermeleobas/224/orig 2025-09-07T07:36:03.9447101Z * [new branch] gh/guilhermeleobas/225/base -> origin/gh/guilhermeleobas/225/base 2025-09-07T07:36:03.9448239Z * [new branch] gh/guilhermeleobas/225/head -> origin/gh/guilhermeleobas/225/head 2025-09-07T07:36:03.9449428Z * [new branch] gh/guilhermeleobas/225/orig -> origin/gh/guilhermeleobas/225/orig 2025-09-07T07:36:03.9451537Z * [new branch] gh/guilhermeleobas/226/base -> origin/gh/guilhermeleobas/226/base 2025-09-07T07:36:03.9452731Z * [new branch] gh/guilhermeleobas/226/head -> origin/gh/guilhermeleobas/226/head 2025-09-07T07:36:03.9453939Z * [new branch] gh/guilhermeleobas/226/orig -> origin/gh/guilhermeleobas/226/orig 2025-09-07T07:36:03.9455787Z * [new branch] gh/guilhermeleobas/227/base -> origin/gh/guilhermeleobas/227/base 2025-09-07T07:36:03.9457096Z * [new branch] gh/guilhermeleobas/227/head -> origin/gh/guilhermeleobas/227/head 2025-09-07T07:36:03.9458859Z * [new branch] gh/guilhermeleobas/227/orig -> origin/gh/guilhermeleobas/227/orig 2025-09-07T07:36:03.9460689Z * [new branch] gh/guilhermeleobas/228/base -> origin/gh/guilhermeleobas/228/base 2025-09-07T07:36:03.9461896Z * [new branch] gh/guilhermeleobas/228/head -> origin/gh/guilhermeleobas/228/head 2025-09-07T07:36:03.9463529Z * [new branch] gh/guilhermeleobas/228/orig -> origin/gh/guilhermeleobas/228/orig 2025-09-07T07:36:03.9465354Z * [new branch] gh/guilhermeleobas/229/base -> origin/gh/guilhermeleobas/229/base 2025-09-07T07:36:03.9466524Z * [new branch] gh/guilhermeleobas/229/head -> origin/gh/guilhermeleobas/229/head 2025-09-07T07:36:03.9467747Z * [new branch] gh/guilhermeleobas/229/orig -> origin/gh/guilhermeleobas/229/orig 2025-09-07T07:36:03.9469503Z * [new branch] gh/guilhermeleobas/230/base -> origin/gh/guilhermeleobas/230/base 2025-09-07T07:36:03.9470697Z * [new branch] gh/guilhermeleobas/230/head -> origin/gh/guilhermeleobas/230/head 2025-09-07T07:36:03.9471867Z * [new branch] gh/guilhermeleobas/230/orig -> origin/gh/guilhermeleobas/230/orig 2025-09-07T07:36:03.9473672Z * [new branch] gh/guilhermeleobas/231/base -> origin/gh/guilhermeleobas/231/base 2025-09-07T07:36:03.9474851Z * [new branch] gh/guilhermeleobas/231/head -> origin/gh/guilhermeleobas/231/head 2025-09-07T07:36:03.9476064Z * [new branch] gh/guilhermeleobas/231/orig -> origin/gh/guilhermeleobas/231/orig 2025-09-07T07:36:03.9477805Z * [new branch] gh/guilhermeleobas/232/base -> origin/gh/guilhermeleobas/232/base 2025-09-07T07:36:03.9479002Z * [new branch] gh/guilhermeleobas/232/head -> origin/gh/guilhermeleobas/232/head 2025-09-07T07:36:03.9480275Z * [new branch] gh/guilhermeleobas/232/orig -> origin/gh/guilhermeleobas/232/orig 2025-09-07T07:36:03.9482092Z * [new branch] gh/guilhermeleobas/233/base -> origin/gh/guilhermeleobas/233/base 2025-09-07T07:36:03.9483160Z * [new branch] gh/guilhermeleobas/233/head -> origin/gh/guilhermeleobas/233/head 2025-09-07T07:36:03.9484362Z * [new branch] gh/guilhermeleobas/233/orig -> origin/gh/guilhermeleobas/233/orig 2025-09-07T07:36:03.9486135Z * [new branch] gh/guilhermeleobas/234/base -> origin/gh/guilhermeleobas/234/base 2025-09-07T07:36:03.9487365Z * [new branch] gh/guilhermeleobas/234/head -> origin/gh/guilhermeleobas/234/head 2025-09-07T07:36:03.9488939Z * [new branch] gh/guilhermeleobas/234/orig -> origin/gh/guilhermeleobas/234/orig 2025-09-07T07:36:03.9490737Z * [new branch] gh/guilhermeleobas/235/base -> origin/gh/guilhermeleobas/235/base 2025-09-07T07:36:03.9491927Z * [new branch] gh/guilhermeleobas/235/head -> origin/gh/guilhermeleobas/235/head 2025-09-07T07:36:03.9493300Z * [new branch] gh/guilhermeleobas/235/orig -> origin/gh/guilhermeleobas/235/orig 2025-09-07T07:36:03.9495053Z * [new branch] gh/guilhermeleobas/236/base -> origin/gh/guilhermeleobas/236/base 2025-09-07T07:36:03.9496205Z * [new branch] gh/guilhermeleobas/236/head -> origin/gh/guilhermeleobas/236/head 2025-09-07T07:36:03.9497407Z * [new branch] gh/guilhermeleobas/236/orig -> origin/gh/guilhermeleobas/236/orig 2025-09-07T07:36:03.9499297Z * [new branch] gh/guilhermeleobas/237/base -> origin/gh/guilhermeleobas/237/base 2025-09-07T07:36:03.9500484Z * [new branch] gh/guilhermeleobas/237/head -> origin/gh/guilhermeleobas/237/head 2025-09-07T07:36:03.9501664Z * [new branch] gh/guilhermeleobas/237/orig -> origin/gh/guilhermeleobas/237/orig 2025-09-07T07:36:03.9503588Z * [new branch] gh/guilhermeleobas/238/base -> origin/gh/guilhermeleobas/238/base 2025-09-07T07:36:03.9504850Z * [new branch] gh/guilhermeleobas/238/head -> origin/gh/guilhermeleobas/238/head 2025-09-07T07:36:03.9506029Z * [new branch] gh/guilhermeleobas/238/orig -> origin/gh/guilhermeleobas/238/orig 2025-09-07T07:36:03.9507850Z * [new branch] gh/guilhermeleobas/239/base -> origin/gh/guilhermeleobas/239/base 2025-09-07T07:36:03.9509018Z * [new branch] gh/guilhermeleobas/239/head -> origin/gh/guilhermeleobas/239/head 2025-09-07T07:36:03.9510292Z * [new branch] gh/guilhermeleobas/239/orig -> origin/gh/guilhermeleobas/239/orig 2025-09-07T07:36:03.9512128Z * [new branch] gh/guilhermeleobas/240/base -> origin/gh/guilhermeleobas/240/base 2025-09-07T07:36:03.9513349Z * [new branch] gh/guilhermeleobas/240/head -> origin/gh/guilhermeleobas/240/head 2025-09-07T07:36:03.9514550Z * [new branch] gh/guilhermeleobas/240/orig -> origin/gh/guilhermeleobas/240/orig 2025-09-07T07:36:03.9516337Z * [new branch] gh/guilhermeleobas/241/base -> origin/gh/guilhermeleobas/241/base 2025-09-07T07:36:03.9517602Z * [new branch] gh/guilhermeleobas/241/head -> origin/gh/guilhermeleobas/241/head 2025-09-07T07:36:03.9518790Z * [new branch] gh/guilhermeleobas/241/orig -> origin/gh/guilhermeleobas/241/orig 2025-09-07T07:36:03.9520547Z * [new branch] gh/guilhermeleobas/242/base -> origin/gh/guilhermeleobas/242/base 2025-09-07T07:36:03.9521809Z * [new branch] gh/guilhermeleobas/242/head -> origin/gh/guilhermeleobas/242/head 2025-09-07T07:36:03.9522979Z * [new branch] gh/guilhermeleobas/242/orig -> origin/gh/guilhermeleobas/242/orig 2025-09-07T07:36:03.9524703Z * [new branch] gh/guilhermeleobas/243/base -> origin/gh/guilhermeleobas/243/base 2025-09-07T07:36:03.9525932Z * [new branch] gh/guilhermeleobas/243/head -> origin/gh/guilhermeleobas/243/head 2025-09-07T07:36:03.9527117Z * [new branch] gh/guilhermeleobas/243/orig -> origin/gh/guilhermeleobas/243/orig 2025-09-07T07:36:03.9529167Z * [new branch] gh/guilhermeleobas/244/base -> origin/gh/guilhermeleobas/244/base 2025-09-07T07:36:03.9530241Z * [new branch] gh/guilhermeleobas/244/head -> origin/gh/guilhermeleobas/244/head 2025-09-07T07:36:03.9531415Z * [new branch] gh/guilhermeleobas/244/orig -> origin/gh/guilhermeleobas/244/orig 2025-09-07T07:36:03.9533129Z * [new branch] gh/guilhermeleobas/245/base -> origin/gh/guilhermeleobas/245/base 2025-09-07T07:36:03.9534340Z * [new branch] gh/guilhermeleobas/245/head -> origin/gh/guilhermeleobas/245/head 2025-09-07T07:36:03.9535513Z * [new branch] gh/guilhermeleobas/245/orig -> origin/gh/guilhermeleobas/245/orig 2025-09-07T07:36:03.9537281Z * [new branch] gh/guilhermeleobas/73/base -> origin/gh/guilhermeleobas/73/base 2025-09-07T07:36:03.9538596Z * [new branch] gh/guilhermeleobas/73/head -> origin/gh/guilhermeleobas/73/head 2025-09-07T07:36:03.9539800Z * [new branch] gh/guilhermeleobas/73/orig -> origin/gh/guilhermeleobas/73/orig 2025-09-07T07:36:03.9541992Z * [new branch] gh/henrylhtsang/140/base -> origin/gh/henrylhtsang/140/base 2025-09-07T07:36:03.9543230Z * [new branch] gh/henrylhtsang/140/head -> origin/gh/henrylhtsang/140/head 2025-09-07T07:36:03.9544434Z * [new branch] gh/henrylhtsang/140/orig -> origin/gh/henrylhtsang/140/orig 2025-09-07T07:36:03.9546068Z * [new branch] gh/henrylhtsang/141/base -> origin/gh/henrylhtsang/141/base 2025-09-07T07:36:03.9547273Z * [new branch] gh/henrylhtsang/141/head -> origin/gh/henrylhtsang/141/head 2025-09-07T07:36:03.9548524Z * [new branch] gh/henrylhtsang/141/orig -> origin/gh/henrylhtsang/141/orig 2025-09-07T07:36:03.9550432Z * [new branch] gh/henrylhtsang/142/base -> origin/gh/henrylhtsang/142/base 2025-09-07T07:36:03.9551731Z * [new branch] gh/henrylhtsang/142/head -> origin/gh/henrylhtsang/142/head 2025-09-07T07:36:03.9552937Z * [new branch] gh/henrylhtsang/142/orig -> origin/gh/henrylhtsang/142/orig 2025-09-07T07:36:03.9554620Z * [new branch] gh/henrylhtsang/143/base -> origin/gh/henrylhtsang/143/base 2025-09-07T07:36:03.9555812Z * [new branch] gh/henrylhtsang/143/head -> origin/gh/henrylhtsang/143/head 2025-09-07T07:36:03.9557035Z * [new branch] gh/henrylhtsang/143/orig -> origin/gh/henrylhtsang/143/orig 2025-09-07T07:36:03.9558791Z * [new branch] gh/henrylhtsang/144/base -> origin/gh/henrylhtsang/144/base 2025-09-07T07:36:03.9559974Z * [new branch] gh/henrylhtsang/144/head -> origin/gh/henrylhtsang/144/head 2025-09-07T07:36:03.9561160Z * [new branch] gh/henrylhtsang/144/orig -> origin/gh/henrylhtsang/144/orig 2025-09-07T07:36:03.9563005Z * [new branch] gh/henrylhtsang/145/base -> origin/gh/henrylhtsang/145/base 2025-09-07T07:36:03.9564314Z * [new branch] gh/henrylhtsang/145/head -> origin/gh/henrylhtsang/145/head 2025-09-07T07:36:03.9565493Z * [new branch] gh/henrylhtsang/145/orig -> origin/gh/henrylhtsang/145/orig 2025-09-07T07:36:03.9567265Z * [new branch] gh/henrylhtsang/146/base -> origin/gh/henrylhtsang/146/base 2025-09-07T07:36:03.9568493Z * [new branch] gh/henrylhtsang/146/head -> origin/gh/henrylhtsang/146/head 2025-09-07T07:36:03.9569680Z * [new branch] gh/henrylhtsang/146/orig -> origin/gh/henrylhtsang/146/orig 2025-09-07T07:36:03.9571405Z * [new branch] gh/henrylhtsang/147/base -> origin/gh/henrylhtsang/147/base 2025-09-07T07:36:03.9572597Z * [new branch] gh/henrylhtsang/147/head -> origin/gh/henrylhtsang/147/head 2025-09-07T07:36:03.9573816Z * [new branch] gh/henrylhtsang/147/orig -> origin/gh/henrylhtsang/147/orig 2025-09-07T07:36:03.9575689Z * [new branch] gh/henrylhtsang/148/base -> origin/gh/henrylhtsang/148/base 2025-09-07T07:36:03.9577593Z * [new branch] gh/henrylhtsang/148/head -> origin/gh/henrylhtsang/148/head 2025-09-07T07:36:03.9578955Z * [new branch] gh/henrylhtsang/148/orig -> origin/gh/henrylhtsang/148/orig 2025-09-07T07:36:03.9580624Z * [new branch] gh/henrylhtsang/149/base -> origin/gh/henrylhtsang/149/base 2025-09-07T07:36:03.9581963Z * [new branch] gh/henrylhtsang/149/head -> origin/gh/henrylhtsang/149/head 2025-09-07T07:36:03.9583243Z * [new branch] gh/henrylhtsang/149/orig -> origin/gh/henrylhtsang/149/orig 2025-09-07T07:36:03.9585582Z * [new branch] gh/huydhn/1/next -> origin/gh/huydhn/1/next 2025-09-07T07:36:03.9587209Z * [new branch] gh/huydhn/2/next -> origin/gh/huydhn/2/next 2025-09-07T07:36:03.9588890Z * [new branch] gh/huydhn/3/next -> origin/gh/huydhn/3/next 2025-09-07T07:36:03.9590699Z * [new branch] gh/huydhn/4/next -> origin/gh/huydhn/4/next 2025-09-07T07:36:03.9592352Z * [new branch] gh/huydhn/5/next -> origin/gh/huydhn/5/next 2025-09-07T07:36:03.9594081Z * [new branch] gh/huydhn/6/next -> origin/gh/huydhn/6/next 2025-09-07T07:36:03.9596278Z * [new branch] gh/int3/97/base -> origin/gh/int3/97/base 2025-09-07T07:36:03.9597473Z * [new branch] gh/int3/97/head -> origin/gh/int3/97/head 2025-09-07T07:36:03.9600152Z * [new branch] gh/isuruf/101/base -> origin/gh/isuruf/101/base 2025-09-07T07:36:03.9601352Z * [new branch] gh/isuruf/101/head -> origin/gh/isuruf/101/head 2025-09-07T07:36:03.9603028Z * [new branch] gh/isuruf/141/base -> origin/gh/isuruf/141/base 2025-09-07T07:36:03.9604630Z * [new branch] gh/isuruf/141/head -> origin/gh/isuruf/141/head 2025-09-07T07:36:03.9605830Z * [new branch] gh/isuruf/141/orig -> origin/gh/isuruf/141/orig 2025-09-07T07:36:03.9607652Z * [new branch] gh/isuruf/142/base -> origin/gh/isuruf/142/base 2025-09-07T07:36:03.9608835Z * [new branch] gh/isuruf/142/head -> origin/gh/isuruf/142/head 2025-09-07T07:36:03.9610045Z * [new branch] gh/isuruf/142/orig -> origin/gh/isuruf/142/orig 2025-09-07T07:36:03.9611701Z * [new branch] gh/isuruf/143/base -> origin/gh/isuruf/143/base 2025-09-07T07:36:03.9612858Z * [new branch] gh/isuruf/143/head -> origin/gh/isuruf/143/head 2025-09-07T07:36:03.9614037Z * [new branch] gh/isuruf/143/orig -> origin/gh/isuruf/143/orig 2025-09-07T07:36:03.9615692Z * [new branch] gh/isuruf/144/base -> origin/gh/isuruf/144/base 2025-09-07T07:36:03.9616853Z * [new branch] gh/isuruf/144/head -> origin/gh/isuruf/144/head 2025-09-07T07:36:03.9618103Z * [new branch] gh/isuruf/144/orig -> origin/gh/isuruf/144/orig 2025-09-07T07:36:03.9619805Z * [new branch] gh/isuruf/145/base -> origin/gh/isuruf/145/base 2025-09-07T07:36:03.9621060Z * [new branch] gh/isuruf/145/head -> origin/gh/isuruf/145/head 2025-09-07T07:36:03.9622311Z * [new branch] gh/isuruf/145/orig -> origin/gh/isuruf/145/orig 2025-09-07T07:36:03.9623992Z * [new branch] gh/isuruf/146/base -> origin/gh/isuruf/146/base 2025-09-07T07:36:03.9625195Z * [new branch] gh/isuruf/146/head -> origin/gh/isuruf/146/head 2025-09-07T07:36:03.9626401Z * [new branch] gh/isuruf/146/orig -> origin/gh/isuruf/146/orig 2025-09-07T07:36:03.9628118Z * [new branch] gh/isuruf/81/base -> origin/gh/isuruf/81/base 2025-09-07T07:36:03.9629326Z * [new branch] gh/isuruf/81/head -> origin/gh/isuruf/81/head 2025-09-07T07:36:03.9630581Z * [new branch] gh/isuruf/81/orig -> origin/gh/isuruf/81/orig 2025-09-07T07:36:03.9632620Z * [new branch] gh/jamesjwu/150/base -> origin/gh/jamesjwu/150/base 2025-09-07T07:36:03.9633761Z * [new branch] gh/jamesjwu/150/head -> origin/gh/jamesjwu/150/head 2025-09-07T07:36:03.9634929Z * [new branch] gh/jamesjwu/150/orig -> origin/gh/jamesjwu/150/orig 2025-09-07T07:36:03.9636818Z * [new branch] gh/jamesjwu/154/base -> origin/gh/jamesjwu/154/base 2025-09-07T07:36:03.9637955Z * [new branch] gh/jamesjwu/154/head -> origin/gh/jamesjwu/154/head 2025-09-07T07:36:03.9639251Z * [new branch] gh/jamesjwu/154/orig -> origin/gh/jamesjwu/154/orig 2025-09-07T07:36:03.9640962Z * [new branch] gh/jamesjwu/155/base -> origin/gh/jamesjwu/155/base 2025-09-07T07:36:03.9642138Z * [new branch] gh/jamesjwu/155/head -> origin/gh/jamesjwu/155/head 2025-09-07T07:36:03.9643331Z * [new branch] gh/jamesjwu/155/orig -> origin/gh/jamesjwu/155/orig 2025-09-07T07:36:03.9645030Z * [new branch] gh/jamesjwu/159/base -> origin/gh/jamesjwu/159/base 2025-09-07T07:36:03.9646213Z * [new branch] gh/jamesjwu/159/head -> origin/gh/jamesjwu/159/head 2025-09-07T07:36:03.9647513Z * [new branch] gh/jamesjwu/159/orig -> origin/gh/jamesjwu/159/orig 2025-09-07T07:36:03.9649457Z * [new branch] gh/jamesjwu/163/base -> origin/gh/jamesjwu/163/base 2025-09-07T07:36:03.9650672Z * [new branch] gh/jamesjwu/163/head -> origin/gh/jamesjwu/163/head 2025-09-07T07:36:03.9651831Z * [new branch] gh/jamesjwu/163/orig -> origin/gh/jamesjwu/163/orig 2025-09-07T07:36:03.9654057Z * [new branch] gh/jamesjwu/171/base -> origin/gh/jamesjwu/171/base 2025-09-07T07:36:03.9655274Z * [new branch] gh/jamesjwu/171/head -> origin/gh/jamesjwu/171/head 2025-09-07T07:36:03.9656422Z * [new branch] gh/jamesjwu/171/orig -> origin/gh/jamesjwu/171/orig 2025-09-07T07:36:03.9658200Z * [new branch] gh/jamesjwu/176/base -> origin/gh/jamesjwu/176/base 2025-09-07T07:36:03.9659418Z * [new branch] gh/jamesjwu/176/head -> origin/gh/jamesjwu/176/head 2025-09-07T07:36:03.9660631Z * [new branch] gh/jamesjwu/176/orig -> origin/gh/jamesjwu/176/orig 2025-09-07T07:36:03.9662364Z * [new branch] gh/jamesjwu/181/base -> origin/gh/jamesjwu/181/base 2025-09-07T07:36:03.9663537Z * [new branch] gh/jamesjwu/181/head -> origin/gh/jamesjwu/181/head 2025-09-07T07:36:03.9664704Z * [new branch] gh/jamesjwu/181/orig -> origin/gh/jamesjwu/181/orig 2025-09-07T07:36:03.9666806Z * [new branch] gh/jamesjwu/182/base -> origin/gh/jamesjwu/182/base 2025-09-07T07:36:03.9668020Z * [new branch] gh/jamesjwu/182/head -> origin/gh/jamesjwu/182/head 2025-09-07T07:36:03.9669220Z * [new branch] gh/jamesjwu/182/orig -> origin/gh/jamesjwu/182/orig 2025-09-07T07:36:03.9670994Z * [new branch] gh/jamesjwu/183/base -> origin/gh/jamesjwu/183/base 2025-09-07T07:36:03.9672615Z * [new branch] gh/jamesjwu/183/head -> origin/gh/jamesjwu/183/head 2025-09-07T07:36:03.9673839Z * [new branch] gh/jamesjwu/183/orig -> origin/gh/jamesjwu/183/orig 2025-09-07T07:36:03.9675792Z * [new branch] gh/jamesjwu/184/base -> origin/gh/jamesjwu/184/base 2025-09-07T07:36:03.9676876Z * [new branch] gh/jamesjwu/184/head -> origin/gh/jamesjwu/184/head 2025-09-07T07:36:03.9678082Z * [new branch] gh/jamesjwu/184/orig -> origin/gh/jamesjwu/184/orig 2025-09-07T07:36:03.9679778Z * [new branch] gh/jamesjwu/185/base -> origin/gh/jamesjwu/185/base 2025-09-07T07:36:03.9681056Z * [new branch] gh/jamesjwu/185/head -> origin/gh/jamesjwu/185/head 2025-09-07T07:36:03.9682239Z * [new branch] gh/jamesjwu/185/orig -> origin/gh/jamesjwu/185/orig 2025-09-07T07:36:03.9683911Z * [new branch] gh/jamesjwu/186/base -> origin/gh/jamesjwu/186/base 2025-09-07T07:36:03.9685093Z * [new branch] gh/jamesjwu/186/head -> origin/gh/jamesjwu/186/head 2025-09-07T07:36:03.9686434Z * [new branch] gh/jamesjwu/186/orig -> origin/gh/jamesjwu/186/orig 2025-09-07T07:36:03.9688291Z * [new branch] gh/jamesjwu/187/base -> origin/gh/jamesjwu/187/base 2025-09-07T07:36:03.9689695Z * [new branch] gh/jamesjwu/187/head -> origin/gh/jamesjwu/187/head 2025-09-07T07:36:03.9690874Z * [new branch] gh/jamesjwu/187/orig -> origin/gh/jamesjwu/187/orig 2025-09-07T07:36:03.9692678Z * [new branch] gh/jamesjwu/188/base -> origin/gh/jamesjwu/188/base 2025-09-07T07:36:03.9693937Z * [new branch] gh/jamesjwu/188/head -> origin/gh/jamesjwu/188/head 2025-09-07T07:36:03.9695128Z * [new branch] gh/jamesjwu/188/orig -> origin/gh/jamesjwu/188/orig 2025-09-07T07:36:03.9697404Z * [new branch] gh/jamesjwu/189/base -> origin/gh/jamesjwu/189/base 2025-09-07T07:36:03.9698760Z * [new branch] gh/jamesjwu/189/head -> origin/gh/jamesjwu/189/head 2025-09-07T07:36:03.9699969Z * [new branch] gh/jamesjwu/189/orig -> origin/gh/jamesjwu/189/orig 2025-09-07T07:36:03.9702150Z * [new branch] gh/jamesjwu/190/base -> origin/gh/jamesjwu/190/base 2025-09-07T07:36:03.9703486Z * [new branch] gh/jamesjwu/190/head -> origin/gh/jamesjwu/190/head 2025-09-07T07:36:03.9706997Z * [new branch] gh/jamesjwu/190/orig -> origin/gh/jamesjwu/190/orig 2025-09-07T07:36:03.9708945Z * [new branch] gh/jamesjwu/52/base -> origin/gh/jamesjwu/52/base 2025-09-07T07:36:03.9710137Z * [new branch] gh/jamesjwu/52/head -> origin/gh/jamesjwu/52/head 2025-09-07T07:36:03.9711800Z * [new branch] gh/jamesjwu/53/base -> origin/gh/jamesjwu/53/base 2025-09-07T07:36:03.9712966Z * [new branch] gh/jamesjwu/53/head -> origin/gh/jamesjwu/53/head 2025-09-07T07:36:03.9714625Z * [new branch] gh/jamesjwu/54/base -> origin/gh/jamesjwu/54/base 2025-09-07T07:36:03.9715997Z * [new branch] gh/jamesjwu/54/head -> origin/gh/jamesjwu/54/head 2025-09-07T07:36:03.9717642Z * [new branch] gh/jamesjwu/55/base -> origin/gh/jamesjwu/55/base 2025-09-07T07:36:03.9718790Z * [new branch] gh/jamesjwu/55/head -> origin/gh/jamesjwu/55/head 2025-09-07T07:36:03.9720437Z * [new branch] gh/jamesjwu/56/base -> origin/gh/jamesjwu/56/base 2025-09-07T07:36:03.9721570Z * [new branch] gh/jamesjwu/56/head -> origin/gh/jamesjwu/56/head 2025-09-07T07:36:03.9723207Z * [new branch] gh/jamesjwu/57/base -> origin/gh/jamesjwu/57/base 2025-09-07T07:36:03.9724445Z * [new branch] gh/jamesjwu/57/head -> origin/gh/jamesjwu/57/head 2025-09-07T07:36:03.9726083Z * [new branch] gh/jamesjwu/58/base -> origin/gh/jamesjwu/58/base 2025-09-07T07:36:03.9727266Z * [new branch] gh/jamesjwu/58/head -> origin/gh/jamesjwu/58/head 2025-09-07T07:36:03.9728915Z * [new branch] gh/jamesjwu/59/base -> origin/gh/jamesjwu/59/base 2025-09-07T07:36:03.9730089Z * [new branch] gh/jamesjwu/59/head -> origin/gh/jamesjwu/59/head 2025-09-07T07:36:03.9731687Z * [new branch] gh/jamesjwu/60/base -> origin/gh/jamesjwu/60/base 2025-09-07T07:36:03.9732834Z * [new branch] gh/jamesjwu/60/head -> origin/gh/jamesjwu/60/head 2025-09-07T07:36:03.9734565Z * [new branch] gh/jamesjwu/61/base -> origin/gh/jamesjwu/61/base 2025-09-07T07:36:03.9735805Z * [new branch] gh/jamesjwu/61/head -> origin/gh/jamesjwu/61/head 2025-09-07T07:36:03.9737416Z * [new branch] gh/jamesjwu/62/base -> origin/gh/jamesjwu/62/base 2025-09-07T07:36:03.9738653Z * [new branch] gh/jamesjwu/62/head -> origin/gh/jamesjwu/62/head 2025-09-07T07:36:03.9740352Z * [new branch] gh/jamesjwu/63/base -> origin/gh/jamesjwu/63/base 2025-09-07T07:36:03.9741553Z * [new branch] gh/jamesjwu/63/head -> origin/gh/jamesjwu/63/head 2025-09-07T07:36:03.9743467Z * [new branch] gh/jamesjwu/64/base -> origin/gh/jamesjwu/64/base 2025-09-07T07:36:03.9744657Z * [new branch] gh/jamesjwu/64/head -> origin/gh/jamesjwu/64/head 2025-09-07T07:36:03.9746284Z * [new branch] gh/jamesjwu/65/base -> origin/gh/jamesjwu/65/base 2025-09-07T07:36:03.9747481Z * [new branch] gh/jamesjwu/65/head -> origin/gh/jamesjwu/65/head 2025-09-07T07:36:03.9749614Z * [new branch] gh/janeyx99/165/base -> origin/gh/janeyx99/165/base 2025-09-07T07:36:03.9750839Z * [new branch] gh/janeyx99/165/head -> origin/gh/janeyx99/165/head 2025-09-07T07:36:03.9752146Z * [new branch] gh/janeyx99/165/orig -> origin/gh/janeyx99/165/orig 2025-09-07T07:36:03.9753742Z * [new branch] gh/janeyx99/201/base -> origin/gh/janeyx99/201/base 2025-09-07T07:36:03.9755029Z * [new branch] gh/janeyx99/201/head -> origin/gh/janeyx99/201/head 2025-09-07T07:36:03.9756208Z * [new branch] gh/janeyx99/201/orig -> origin/gh/janeyx99/201/orig 2025-09-07T07:36:03.9758270Z * [new branch] gh/janeyx99/225/base -> origin/gh/janeyx99/225/base 2025-09-07T07:36:03.9759482Z * [new branch] gh/janeyx99/225/head -> origin/gh/janeyx99/225/head 2025-09-07T07:36:03.9760676Z * [new branch] gh/janeyx99/225/orig -> origin/gh/janeyx99/225/orig 2025-09-07T07:36:03.9762391Z * [new branch] gh/janeyx99/296/base -> origin/gh/janeyx99/296/base 2025-09-07T07:36:03.9763574Z * [new branch] gh/janeyx99/296/head -> origin/gh/janeyx99/296/head 2025-09-07T07:36:03.9764786Z * [new branch] gh/janeyx99/296/orig -> origin/gh/janeyx99/296/orig 2025-09-07T07:36:03.9766458Z * [new branch] gh/janeyx99/297/base -> origin/gh/janeyx99/297/base 2025-09-07T07:36:03.9767652Z * [new branch] gh/janeyx99/297/head -> origin/gh/janeyx99/297/head 2025-09-07T07:36:03.9768820Z * [new branch] gh/janeyx99/297/orig -> origin/gh/janeyx99/297/orig 2025-09-07T07:36:03.9770972Z * [new branch] gh/janeyx99/298/base -> origin/gh/janeyx99/298/base 2025-09-07T07:36:03.9772244Z * [new branch] gh/janeyx99/298/head -> origin/gh/janeyx99/298/head 2025-09-07T07:36:03.9773444Z * [new branch] gh/janeyx99/298/orig -> origin/gh/janeyx99/298/orig 2025-09-07T07:36:03.9775197Z * [new branch] gh/janeyx99/299/base -> origin/gh/janeyx99/299/base 2025-09-07T07:36:03.9776430Z * [new branch] gh/janeyx99/299/head -> origin/gh/janeyx99/299/head 2025-09-07T07:36:03.9777734Z * [new branch] gh/janeyx99/299/orig -> origin/gh/janeyx99/299/orig 2025-09-07T07:36:03.9779584Z * [new branch] gh/janeyx99/300/base -> origin/gh/janeyx99/300/base 2025-09-07T07:36:03.9780936Z * [new branch] gh/janeyx99/300/head -> origin/gh/janeyx99/300/head 2025-09-07T07:36:03.9782122Z * [new branch] gh/janeyx99/300/orig -> origin/gh/janeyx99/300/orig 2025-09-07T07:36:03.9784396Z * [new branch] gh/janeyx99/301/base -> origin/gh/janeyx99/301/base 2025-09-07T07:36:03.9785745Z * [new branch] gh/janeyx99/301/head -> origin/gh/janeyx99/301/head 2025-09-07T07:36:03.9786899Z * [new branch] gh/janeyx99/301/orig -> origin/gh/janeyx99/301/orig 2025-09-07T07:36:03.9788675Z * [new branch] gh/janeyx99/302/base -> origin/gh/janeyx99/302/base 2025-09-07T07:36:03.9789981Z * [new branch] gh/janeyx99/302/head -> origin/gh/janeyx99/302/head 2025-09-07T07:36:03.9791785Z * [new branch] gh/janeyx99/303/base -> origin/gh/janeyx99/303/base 2025-09-07T07:36:03.9792948Z * [new branch] gh/janeyx99/303/head -> origin/gh/janeyx99/303/head 2025-09-07T07:36:03.9794941Z * [new branch] gh/janeyx99/88/base -> origin/gh/janeyx99/88/base 2025-09-07T07:36:03.9796159Z * [new branch] gh/janeyx99/88/head -> origin/gh/janeyx99/88/head 2025-09-07T07:36:03.9797328Z * [new branch] gh/janeyx99/88/orig -> origin/gh/janeyx99/88/orig 2025-09-07T07:36:03.9799499Z * [new branch] gh/jansel/360/base -> origin/gh/jansel/360/base 2025-09-07T07:36:03.9800664Z * [new branch] gh/jansel/360/head -> origin/gh/jansel/360/head 2025-09-07T07:36:03.9802383Z * [new branch] gh/jansel/451/base -> origin/gh/jansel/451/base 2025-09-07T07:36:03.9803768Z * [new branch] gh/jansel/451/head -> origin/gh/jansel/451/head 2025-09-07T07:36:03.9805002Z * [new branch] gh/jansel/451/orig -> origin/gh/jansel/451/orig 2025-09-07T07:36:03.9806809Z * [new branch] gh/jansel/462/base -> origin/gh/jansel/462/base 2025-09-07T07:36:03.9807984Z * [new branch] gh/jansel/462/head -> origin/gh/jansel/462/head 2025-09-07T07:36:03.9809138Z * [new branch] gh/jansel/462/orig -> origin/gh/jansel/462/orig 2025-09-07T07:36:03.9810862Z * [new branch] gh/jansel/531/base -> origin/gh/jansel/531/base 2025-09-07T07:36:03.9812049Z * [new branch] gh/jansel/531/head -> origin/gh/jansel/531/head 2025-09-07T07:36:03.9813246Z * [new branch] gh/jansel/531/orig -> origin/gh/jansel/531/orig 2025-09-07T07:36:03.9815427Z * [new branch] gh/jbschlosser/208/head -> origin/gh/jbschlosser/208/head 2025-09-07T07:36:03.9817193Z * [new branch] gh/jbschlosser/247/base -> origin/gh/jbschlosser/247/base 2025-09-07T07:36:03.9818514Z * [new branch] gh/jbschlosser/247/head -> origin/gh/jbschlosser/247/head 2025-09-07T07:36:03.9819701Z * [new branch] gh/jbschlosser/247/orig -> origin/gh/jbschlosser/247/orig 2025-09-07T07:36:03.9821551Z * [new branch] gh/jbschlosser/248/base -> origin/gh/jbschlosser/248/base 2025-09-07T07:36:03.9822760Z * [new branch] gh/jbschlosser/248/head -> origin/gh/jbschlosser/248/head 2025-09-07T07:36:03.9823932Z * [new branch] gh/jbschlosser/248/orig -> origin/gh/jbschlosser/248/orig 2025-09-07T07:36:03.9825901Z * [new branch] gh/jbschlosser/250/base -> origin/gh/jbschlosser/250/base 2025-09-07T07:36:03.9827163Z * [new branch] gh/jbschlosser/250/head -> origin/gh/jbschlosser/250/head 2025-09-07T07:36:03.9828355Z * [new branch] gh/jbschlosser/250/orig -> origin/gh/jbschlosser/250/orig 2025-09-07T07:36:03.9830450Z * [new branch] gh/jiayisunx/59/base -> origin/gh/jiayisunx/59/base 2025-09-07T07:36:03.9831653Z * [new branch] gh/jiayisunx/59/head -> origin/gh/jiayisunx/59/head 2025-09-07T07:36:03.9832850Z * [new branch] gh/jiayisunx/59/orig -> origin/gh/jiayisunx/59/orig 2025-09-07T07:36:03.9834477Z * [new branch] gh/jiayisunx/61/base -> origin/gh/jiayisunx/61/base 2025-09-07T07:36:03.9835676Z * [new branch] gh/jiayisunx/61/head -> origin/gh/jiayisunx/61/head 2025-09-07T07:36:03.9836959Z * [new branch] gh/jiayisunx/61/orig -> origin/gh/jiayisunx/61/orig 2025-09-07T07:36:03.9838623Z * [new branch] gh/jiayisunx/64/base -> origin/gh/jiayisunx/64/base 2025-09-07T07:36:03.9839849Z * [new branch] gh/jiayisunx/64/head -> origin/gh/jiayisunx/64/head 2025-09-07T07:36:03.9841006Z * [new branch] gh/jiayisunx/64/orig -> origin/gh/jiayisunx/64/orig 2025-09-07T07:36:03.9842711Z * [new branch] gh/jiayisunx/65/base -> origin/gh/jiayisunx/65/base 2025-09-07T07:36:03.9843964Z * [new branch] gh/jiayisunx/65/head -> origin/gh/jiayisunx/65/head 2025-09-07T07:36:03.9845168Z * [new branch] gh/jiayisunx/65/orig -> origin/gh/jiayisunx/65/orig 2025-09-07T07:36:03.9846799Z * [new branch] gh/jiayisunx/66/base -> origin/gh/jiayisunx/66/base 2025-09-07T07:36:03.9847991Z * [new branch] gh/jiayisunx/66/head -> origin/gh/jiayisunx/66/head 2025-09-07T07:36:03.9849137Z * [new branch] gh/jiayisunx/66/orig -> origin/gh/jiayisunx/66/orig 2025-09-07T07:36:03.9850810Z * [new branch] gh/jiayisunx/67/base -> origin/gh/jiayisunx/67/base 2025-09-07T07:36:03.9851961Z * [new branch] gh/jiayisunx/67/head -> origin/gh/jiayisunx/67/head 2025-09-07T07:36:03.9853140Z * [new branch] gh/jiayisunx/67/orig -> origin/gh/jiayisunx/67/orig 2025-09-07T07:36:03.9854942Z * [new branch] gh/jiayisunx/68/base -> origin/gh/jiayisunx/68/base 2025-09-07T07:36:03.9856118Z * [new branch] gh/jiayisunx/68/head -> origin/gh/jiayisunx/68/head 2025-09-07T07:36:03.9857288Z * [new branch] gh/jiayisunx/68/orig -> origin/gh/jiayisunx/68/orig 2025-09-07T07:36:03.9859156Z * [new branch] gh/jiayisunx/69/base -> origin/gh/jiayisunx/69/base 2025-09-07T07:36:03.9860459Z * [new branch] gh/jiayisunx/69/head -> origin/gh/jiayisunx/69/head 2025-09-07T07:36:03.9861675Z * [new branch] gh/jiayisunx/69/orig -> origin/gh/jiayisunx/69/orig 2025-09-07T07:36:03.9863413Z * [new branch] gh/jiayisunx/70/base -> origin/gh/jiayisunx/70/base 2025-09-07T07:36:03.9864587Z * [new branch] gh/jiayisunx/70/head -> origin/gh/jiayisunx/70/head 2025-09-07T07:36:03.9865788Z * [new branch] gh/jiayisunx/70/orig -> origin/gh/jiayisunx/70/orig 2025-09-07T07:36:03.9867486Z * [new branch] gh/jiayisunx/71/base -> origin/gh/jiayisunx/71/base 2025-09-07T07:36:03.9868672Z * [new branch] gh/jiayisunx/71/head -> origin/gh/jiayisunx/71/head 2025-09-07T07:36:03.9869853Z * [new branch] gh/jiayisunx/71/orig -> origin/gh/jiayisunx/71/orig 2025-09-07T07:36:03.9871652Z * [new branch] gh/jiayisunx/72/base -> origin/gh/jiayisunx/72/base 2025-09-07T07:36:03.9872842Z * [new branch] gh/jiayisunx/72/head -> origin/gh/jiayisunx/72/head 2025-09-07T07:36:03.9874061Z * [new branch] gh/jiayisunx/72/orig -> origin/gh/jiayisunx/72/orig 2025-09-07T07:36:03.9875780Z * [new branch] gh/jiayisunx/73/base -> origin/gh/jiayisunx/73/base 2025-09-07T07:36:03.9877045Z * [new branch] gh/jiayisunx/73/head -> origin/gh/jiayisunx/73/head 2025-09-07T07:36:03.9878245Z * [new branch] gh/jiayisunx/73/orig -> origin/gh/jiayisunx/73/orig 2025-09-07T07:36:03.9879894Z * [new branch] gh/jiayisunx/74/base -> origin/gh/jiayisunx/74/base 2025-09-07T07:36:03.9881106Z * [new branch] gh/jiayisunx/74/head -> origin/gh/jiayisunx/74/head 2025-09-07T07:36:03.9882276Z * [new branch] gh/jiayisunx/74/orig -> origin/gh/jiayisunx/74/orig 2025-09-07T07:36:03.9884077Z * [new branch] gh/jiayisunx/75/base -> origin/gh/jiayisunx/75/base 2025-09-07T07:36:03.9885194Z * [new branch] gh/jiayisunx/75/head -> origin/gh/jiayisunx/75/head 2025-09-07T07:36:03.9886248Z * [new branch] gh/jiayisunx/75/orig -> origin/gh/jiayisunx/75/orig 2025-09-07T07:36:03.9887931Z * [new branch] gh/jiayisunx/76/base -> origin/gh/jiayisunx/76/base 2025-09-07T07:36:03.9889049Z * [new branch] gh/jiayisunx/76/head -> origin/gh/jiayisunx/76/head 2025-09-07T07:36:03.9890236Z * [new branch] gh/jiayisunx/76/orig -> origin/gh/jiayisunx/76/orig 2025-09-07T07:36:03.9892255Z * [new branch] gh/jjwu@meta.com/1/base -> origin/gh/jjwu@meta.com/1/base 2025-09-07T07:36:03.9893437Z * [new branch] gh/jjwu@meta.com/1/head -> origin/gh/jjwu@meta.com/1/head 2025-09-07T07:36:03.9895525Z * [new branch] gh/justinchuby/111/base -> origin/gh/justinchuby/111/base 2025-09-07T07:36:03.9896940Z * [new branch] gh/justinchuby/111/head -> origin/gh/justinchuby/111/head 2025-09-07T07:36:03.9898264Z * [new branch] gh/justinchuby/111/orig -> origin/gh/justinchuby/111/orig 2025-09-07T07:36:03.9900004Z * [new branch] gh/justinchuby/112/base -> origin/gh/justinchuby/112/base 2025-09-07T07:36:03.9901188Z * [new branch] gh/justinchuby/112/head -> origin/gh/justinchuby/112/head 2025-09-07T07:36:03.9902543Z * [new branch] gh/justinchuby/112/orig -> origin/gh/justinchuby/112/orig 2025-09-07T07:36:03.9904436Z * [new branch] gh/justinchuby/113/base -> origin/gh/justinchuby/113/base 2025-09-07T07:36:03.9905655Z * [new branch] gh/justinchuby/113/head -> origin/gh/justinchuby/113/head 2025-09-07T07:36:03.9906890Z * [new branch] gh/justinchuby/113/orig -> origin/gh/justinchuby/113/orig 2025-09-07T07:36:03.9908461Z * [new branch] gh/justinchuby/114/base -> origin/gh/justinchuby/114/base 2025-09-07T07:36:03.9909708Z * [new branch] gh/justinchuby/114/head -> origin/gh/justinchuby/114/head 2025-09-07T07:36:03.9910873Z * [new branch] gh/justinchuby/114/orig -> origin/gh/justinchuby/114/orig 2025-09-07T07:36:03.9912541Z * [new branch] gh/justinchuby/115/base -> origin/gh/justinchuby/115/base 2025-09-07T07:36:03.9913720Z * [new branch] gh/justinchuby/115/head -> origin/gh/justinchuby/115/head 2025-09-07T07:36:03.9914824Z * [new branch] gh/justinchuby/115/orig -> origin/gh/justinchuby/115/orig 2025-09-07T07:36:03.9916921Z * [new branch] gh/karthickai/1/base -> origin/gh/karthickai/1/base 2025-09-07T07:36:03.9918337Z * [new branch] gh/karthickai/1/head -> origin/gh/karthickai/1/head 2025-09-07T07:36:03.9919496Z * [new branch] gh/karthickai/1/orig -> origin/gh/karthickai/1/orig 2025-09-07T07:36:03.9921183Z * [new branch] gh/karthickai/2/base -> origin/gh/karthickai/2/base 2025-09-07T07:36:03.9922362Z * [new branch] gh/karthickai/2/head -> origin/gh/karthickai/2/head 2025-09-07T07:36:03.9923538Z * [new branch] gh/karthickai/2/orig -> origin/gh/karthickai/2/orig 2025-09-07T07:36:03.9925613Z * [new branch] gh/kurtamohler/32/base -> origin/gh/kurtamohler/32/base 2025-09-07T07:36:03.9926779Z * [new branch] gh/kurtamohler/32/head -> origin/gh/kurtamohler/32/head 2025-09-07T07:36:03.9928033Z * [new branch] gh/kurtamohler/32/orig -> origin/gh/kurtamohler/32/orig 2025-09-07T07:36:03.9929691Z * [new branch] gh/kurtamohler/33/base -> origin/gh/kurtamohler/33/base 2025-09-07T07:36:03.9930916Z * [new branch] gh/kurtamohler/33/head -> origin/gh/kurtamohler/33/head 2025-09-07T07:36:03.9932211Z * [new branch] gh/kurtamohler/33/orig -> origin/gh/kurtamohler/33/orig 2025-09-07T07:36:03.9934031Z * [new branch] gh/kurtamohler/34/base -> origin/gh/kurtamohler/34/base 2025-09-07T07:36:03.9935278Z * [new branch] gh/kurtamohler/34/head -> origin/gh/kurtamohler/34/head 2025-09-07T07:36:03.9936383Z * [new branch] gh/kurtamohler/34/orig -> origin/gh/kurtamohler/34/orig 2025-09-07T07:36:03.9938143Z * [new branch] gh/kurtamohler/41/base -> origin/gh/kurtamohler/41/base 2025-09-07T07:36:03.9939338Z * [new branch] gh/kurtamohler/41/head -> origin/gh/kurtamohler/41/head 2025-09-07T07:36:03.9940509Z * [new branch] gh/kurtamohler/41/orig -> origin/gh/kurtamohler/41/orig 2025-09-07T07:36:03.9942206Z * [new branch] gh/kurtamohler/46/base -> origin/gh/kurtamohler/46/base 2025-09-07T07:36:03.9943411Z * [new branch] gh/kurtamohler/46/head -> origin/gh/kurtamohler/46/head 2025-09-07T07:36:03.9944658Z * [new branch] gh/kurtamohler/46/orig -> origin/gh/kurtamohler/46/orig 2025-09-07T07:36:03.9946363Z * [new branch] gh/kurtamohler/47/base -> origin/gh/kurtamohler/47/base 2025-09-07T07:36:03.9947669Z * [new branch] gh/kurtamohler/47/head -> origin/gh/kurtamohler/47/head 2025-09-07T07:36:03.9948857Z * [new branch] gh/kurtamohler/47/orig -> origin/gh/kurtamohler/47/orig 2025-09-07T07:36:03.9950604Z * [new branch] gh/kurtamohler/48/base -> origin/gh/kurtamohler/48/base 2025-09-07T07:36:03.9951801Z * [new branch] gh/kurtamohler/48/head -> origin/gh/kurtamohler/48/head 2025-09-07T07:36:03.9952975Z * [new branch] gh/kurtamohler/48/orig -> origin/gh/kurtamohler/48/orig 2025-09-07T07:36:03.9954643Z * [new branch] gh/kurtamohler/49/base -> origin/gh/kurtamohler/49/base 2025-09-07T07:36:03.9955841Z * [new branch] gh/kurtamohler/49/head -> origin/gh/kurtamohler/49/head 2025-09-07T07:36:03.9957030Z * [new branch] gh/kurtamohler/49/orig -> origin/gh/kurtamohler/49/orig 2025-09-07T07:36:03.9958716Z * [new branch] gh/kurtamohler/50/base -> origin/gh/kurtamohler/50/base 2025-09-07T07:36:03.9959929Z * [new branch] gh/kurtamohler/50/head -> origin/gh/kurtamohler/50/head 2025-09-07T07:36:03.9961091Z * [new branch] gh/kurtamohler/50/orig -> origin/gh/kurtamohler/50/orig 2025-09-07T07:36:03.9963339Z * [new branch] gh/kwen2501/130/base -> origin/gh/kwen2501/130/base 2025-09-07T07:36:03.9964648Z * [new branch] gh/kwen2501/130/head -> origin/gh/kwen2501/130/head 2025-09-07T07:36:03.9965870Z * [new branch] gh/kwen2501/130/orig -> origin/gh/kwen2501/130/orig 2025-09-07T07:36:03.9967743Z * [new branch] gh/kwen2501/15/base -> origin/gh/kwen2501/15/base 2025-09-07T07:36:03.9968916Z * [new branch] gh/kwen2501/15/head -> origin/gh/kwen2501/15/head 2025-09-07T07:36:03.9970611Z * [new branch] gh/kwen2501/156/base -> origin/gh/kwen2501/156/base 2025-09-07T07:36:03.9971777Z * [new branch] gh/kwen2501/156/head -> origin/gh/kwen2501/156/head 2025-09-07T07:36:03.9972949Z * [new branch] gh/kwen2501/156/orig -> origin/gh/kwen2501/156/orig 2025-09-07T07:36:03.9974678Z * [new branch] gh/kwen2501/170/base -> origin/gh/kwen2501/170/base 2025-09-07T07:36:03.9975851Z * [new branch] gh/kwen2501/170/head -> origin/gh/kwen2501/170/head 2025-09-07T07:36:03.9977597Z * [new branch] gh/kwen2501/186/base -> origin/gh/kwen2501/186/base 2025-09-07T07:36:03.9978894Z * [new branch] gh/kwen2501/186/head -> origin/gh/kwen2501/186/head 2025-09-07T07:36:03.9980078Z * [new branch] gh/kwen2501/186/orig -> origin/gh/kwen2501/186/orig 2025-09-07T07:36:03.9981673Z * [new branch] gh/kwen2501/187/base -> origin/gh/kwen2501/187/base 2025-09-07T07:36:03.9983616Z * [new branch] gh/kwen2501/187/head -> origin/gh/kwen2501/187/head 2025-09-07T07:36:03.9984868Z * [new branch] gh/kwen2501/187/orig -> origin/gh/kwen2501/187/orig 2025-09-07T07:36:03.9986561Z * [new branch] gh/kwen2501/188/base -> origin/gh/kwen2501/188/base 2025-09-07T07:36:03.9987728Z * [new branch] gh/kwen2501/188/head -> origin/gh/kwen2501/188/head 2025-09-07T07:36:03.9988918Z * [new branch] gh/kwen2501/188/orig -> origin/gh/kwen2501/188/orig 2025-09-07T07:36:03.9990683Z * [new branch] gh/kwen2501/194/base -> origin/gh/kwen2501/194/base 2025-09-07T07:36:03.9991863Z * [new branch] gh/kwen2501/194/head -> origin/gh/kwen2501/194/head 2025-09-07T07:36:03.9993088Z * [new branch] gh/kwen2501/194/orig -> origin/gh/kwen2501/194/orig 2025-09-07T07:36:03.9994770Z * [new branch] gh/kwen2501/199/base -> origin/gh/kwen2501/199/base 2025-09-07T07:36:03.9995964Z * [new branch] gh/kwen2501/199/head -> origin/gh/kwen2501/199/head 2025-09-07T07:36:03.9997125Z * [new branch] gh/kwen2501/199/orig -> origin/gh/kwen2501/199/orig 2025-09-07T07:36:03.9998732Z * [new branch] gh/kwen2501/200/base -> origin/gh/kwen2501/200/base 2025-09-07T07:36:04.0000042Z * [new branch] gh/kwen2501/200/head -> origin/gh/kwen2501/200/head 2025-09-07T07:36:04.0001210Z * [new branch] gh/kwen2501/200/orig -> origin/gh/kwen2501/200/orig 2025-09-07T07:36:04.0003007Z * [new branch] gh/kwen2501/201/base -> origin/gh/kwen2501/201/base 2025-09-07T07:36:04.0004396Z * [new branch] gh/kwen2501/201/head -> origin/gh/kwen2501/201/head 2025-09-07T07:36:04.0005599Z * [new branch] gh/kwen2501/201/orig -> origin/gh/kwen2501/201/orig 2025-09-07T07:36:04.0007299Z * [new branch] gh/kwen2501/203/base -> origin/gh/kwen2501/203/base 2025-09-07T07:36:04.0008479Z * [new branch] gh/kwen2501/203/head -> origin/gh/kwen2501/203/head 2025-09-07T07:36:04.0009645Z * [new branch] gh/kwen2501/203/orig -> origin/gh/kwen2501/203/orig 2025-09-07T07:36:04.0011366Z * [new branch] gh/kwen2501/204/base -> origin/gh/kwen2501/204/base 2025-09-07T07:36:04.0012512Z * [new branch] gh/kwen2501/204/head -> origin/gh/kwen2501/204/head 2025-09-07T07:36:04.0013698Z * [new branch] gh/kwen2501/204/orig -> origin/gh/kwen2501/204/orig 2025-09-07T07:36:04.0015355Z * [new branch] gh/kwen2501/205/base -> origin/gh/kwen2501/205/base 2025-09-07T07:36:04.0016585Z * [new branch] gh/kwen2501/205/head -> origin/gh/kwen2501/205/head 2025-09-07T07:36:04.0017774Z * [new branch] gh/kwen2501/205/orig -> origin/gh/kwen2501/205/orig 2025-09-07T07:36:04.0019524Z * [new branch] gh/kwen2501/206/base -> origin/gh/kwen2501/206/base 2025-09-07T07:36:04.0020721Z * [new branch] gh/kwen2501/206/head -> origin/gh/kwen2501/206/head 2025-09-07T07:36:04.0021898Z * [new branch] gh/kwen2501/206/orig -> origin/gh/kwen2501/206/orig 2025-09-07T07:36:04.0023590Z * [new branch] gh/kwen2501/207/base -> origin/gh/kwen2501/207/base 2025-09-07T07:36:04.0024755Z * [new branch] gh/kwen2501/207/head -> origin/gh/kwen2501/207/head 2025-09-07T07:36:04.0026051Z * [new branch] gh/kwen2501/207/orig -> origin/gh/kwen2501/207/orig 2025-09-07T07:36:04.0027739Z * [new branch] gh/kwen2501/208/base -> origin/gh/kwen2501/208/base 2025-09-07T07:36:04.0028923Z * [new branch] gh/kwen2501/208/head -> origin/gh/kwen2501/208/head 2025-09-07T07:36:04.0030098Z * [new branch] gh/kwen2501/208/orig -> origin/gh/kwen2501/208/orig 2025-09-07T07:36:04.0032342Z * [new branch] gh/kwen2501/209/base -> origin/gh/kwen2501/209/base 2025-09-07T07:36:04.0033494Z * [new branch] gh/kwen2501/209/head -> origin/gh/kwen2501/209/head 2025-09-07T07:36:04.0034672Z * [new branch] gh/kwen2501/209/orig -> origin/gh/kwen2501/209/orig 2025-09-07T07:36:04.0036430Z * [new branch] gh/kwen2501/210/base -> origin/gh/kwen2501/210/base 2025-09-07T07:36:04.0037743Z * [new branch] gh/kwen2501/210/head -> origin/gh/kwen2501/210/head 2025-09-07T07:36:04.0038923Z * [new branch] gh/kwen2501/210/orig -> origin/gh/kwen2501/210/orig 2025-09-07T07:36:04.0040650Z * [new branch] gh/kwen2501/211/base -> origin/gh/kwen2501/211/base 2025-09-07T07:36:04.0041848Z * [new branch] gh/kwen2501/211/head -> origin/gh/kwen2501/211/head 2025-09-07T07:36:04.0043631Z * [new branch] gh/kwen2501/212/base -> origin/gh/kwen2501/212/base 2025-09-07T07:36:04.0044827Z * [new branch] gh/kwen2501/212/head -> origin/gh/kwen2501/212/head 2025-09-07T07:36:04.0046008Z * [new branch] gh/kwen2501/212/orig -> origin/gh/kwen2501/212/orig 2025-09-07T07:36:04.0047738Z * [new branch] gh/kwen2501/213/base -> origin/gh/kwen2501/213/base 2025-09-07T07:36:04.0048933Z * [new branch] gh/kwen2501/213/head -> origin/gh/kwen2501/213/head 2025-09-07T07:36:04.0050156Z * [new branch] gh/kwen2501/213/orig -> origin/gh/kwen2501/213/orig 2025-09-07T07:36:04.0051906Z * [new branch] gh/kwen2501/214/base -> origin/gh/kwen2501/214/base 2025-09-07T07:36:04.0053098Z * [new branch] gh/kwen2501/214/head -> origin/gh/kwen2501/214/head 2025-09-07T07:36:04.0054274Z * [new branch] gh/kwen2501/214/orig -> origin/gh/kwen2501/214/orig 2025-09-07T07:36:04.0056050Z * [new branch] gh/kwen2501/215/base -> origin/gh/kwen2501/215/base 2025-09-07T07:36:04.0057259Z * [new branch] gh/kwen2501/215/head -> origin/gh/kwen2501/215/head 2025-09-07T07:36:04.0058539Z * [new branch] gh/kwen2501/215/orig -> origin/gh/kwen2501/215/orig 2025-09-07T07:36:04.0060290Z * [new branch] gh/kwen2501/216/base -> origin/gh/kwen2501/216/base 2025-09-07T07:36:04.0061501Z * [new branch] gh/kwen2501/216/head -> origin/gh/kwen2501/216/head 2025-09-07T07:36:04.0062722Z * [new branch] gh/kwen2501/216/orig -> origin/gh/kwen2501/216/orig 2025-09-07T07:36:04.0064417Z * [new branch] gh/kwen2501/217/base -> origin/gh/kwen2501/217/base 2025-09-07T07:36:04.0065605Z * [new branch] gh/kwen2501/217/head -> origin/gh/kwen2501/217/head 2025-09-07T07:36:04.0066904Z * [new branch] gh/kwen2501/217/orig -> origin/gh/kwen2501/217/orig 2025-09-07T07:36:04.0068616Z * [new branch] gh/kwen2501/218/base -> origin/gh/kwen2501/218/base 2025-09-07T07:36:04.0069848Z * [new branch] gh/kwen2501/218/head -> origin/gh/kwen2501/218/head 2025-09-07T07:36:04.0071012Z * [new branch] gh/kwen2501/218/orig -> origin/gh/kwen2501/218/orig 2025-09-07T07:36:04.0072844Z * [new branch] gh/kwen2501/219/base -> origin/gh/kwen2501/219/base 2025-09-07T07:36:04.0074002Z * [new branch] gh/kwen2501/219/head -> origin/gh/kwen2501/219/head 2025-09-07T07:36:04.0075157Z * [new branch] gh/kwen2501/219/orig -> origin/gh/kwen2501/219/orig 2025-09-07T07:36:04.0076880Z * [new branch] gh/kwen2501/220/base -> origin/gh/kwen2501/220/base 2025-09-07T07:36:04.0078069Z * [new branch] gh/kwen2501/220/head -> origin/gh/kwen2501/220/head 2025-09-07T07:36:04.0079212Z * [new branch] gh/kwen2501/220/orig -> origin/gh/kwen2501/220/orig 2025-09-07T07:36:04.0080998Z * [new branch] gh/kwen2501/221/base -> origin/gh/kwen2501/221/base 2025-09-07T07:36:04.0082630Z * [new branch] gh/kwen2501/221/head -> origin/gh/kwen2501/221/head 2025-09-07T07:36:04.0083909Z * [new branch] gh/kwen2501/221/orig -> origin/gh/kwen2501/221/orig 2025-09-07T07:36:04.0085674Z * [new branch] gh/kwen2501/222/base -> origin/gh/kwen2501/222/base 2025-09-07T07:36:04.0086875Z * [new branch] gh/kwen2501/222/head -> origin/gh/kwen2501/222/head 2025-09-07T07:36:04.0088034Z * [new branch] gh/kwen2501/222/orig -> origin/gh/kwen2501/222/orig 2025-09-07T07:36:04.0089752Z * [new branch] gh/kwen2501/223/base -> origin/gh/kwen2501/223/base 2025-09-07T07:36:04.0090919Z * [new branch] gh/kwen2501/223/head -> origin/gh/kwen2501/223/head 2025-09-07T07:36:04.0092061Z * [new branch] gh/kwen2501/223/orig -> origin/gh/kwen2501/223/orig 2025-09-07T07:36:04.0093756Z * [new branch] gh/kwen2501/224/base -> origin/gh/kwen2501/224/base 2025-09-07T07:36:04.0094909Z * [new branch] gh/kwen2501/224/head -> origin/gh/kwen2501/224/head 2025-09-07T07:36:04.0096079Z * [new branch] gh/kwen2501/224/orig -> origin/gh/kwen2501/224/orig 2025-09-07T07:36:04.0097915Z * [new branch] gh/kwen2501/225/base -> origin/gh/kwen2501/225/base 2025-09-07T07:36:04.0099132Z * [new branch] gh/kwen2501/225/head -> origin/gh/kwen2501/225/head 2025-09-07T07:36:04.0100383Z * [new branch] gh/kwen2501/225/orig -> origin/gh/kwen2501/225/orig 2025-09-07T07:36:04.0102047Z * [new branch] gh/kwen2501/226/base -> origin/gh/kwen2501/226/base 2025-09-07T07:36:04.0103384Z * [new branch] gh/kwen2501/226/head -> origin/gh/kwen2501/226/head 2025-09-07T07:36:04.0107910Z * [new branch] gh/kwen2501/226/orig -> origin/gh/kwen2501/226/orig 2025-09-07T07:36:04.0109560Z * [new branch] gh/kwen2501/227/base -> origin/gh/kwen2501/227/base 2025-09-07T07:36:04.0110940Z * [new branch] gh/kwen2501/227/head -> origin/gh/kwen2501/227/head 2025-09-07T07:36:04.0112151Z * [new branch] gh/kwen2501/227/orig -> origin/gh/kwen2501/227/orig 2025-09-07T07:36:04.0113840Z * [new branch] gh/kwen2501/228/base -> origin/gh/kwen2501/228/base 2025-09-07T07:36:04.0115001Z * [new branch] gh/kwen2501/228/head -> origin/gh/kwen2501/228/head 2025-09-07T07:36:04.0116162Z * [new branch] gh/kwen2501/228/orig -> origin/gh/kwen2501/228/orig 2025-09-07T07:36:04.0117893Z * [new branch] gh/kwen2501/229/base -> origin/gh/kwen2501/229/base 2025-09-07T07:36:04.0119119Z * [new branch] gh/kwen2501/229/head -> origin/gh/kwen2501/229/head 2025-09-07T07:36:04.0120380Z * [new branch] gh/kwen2501/229/orig -> origin/gh/kwen2501/229/orig 2025-09-07T07:36:04.0122125Z * [new branch] gh/kwen2501/230/base -> origin/gh/kwen2501/230/base 2025-09-07T07:36:04.0123292Z * [new branch] gh/kwen2501/230/head -> origin/gh/kwen2501/230/head 2025-09-07T07:36:04.0124484Z * [new branch] gh/kwen2501/230/orig -> origin/gh/kwen2501/230/orig 2025-09-07T07:36:04.0126198Z * [new branch] gh/kwen2501/231/base -> origin/gh/kwen2501/231/base 2025-09-07T07:36:04.0127375Z * [new branch] gh/kwen2501/231/head -> origin/gh/kwen2501/231/head 2025-09-07T07:36:04.0128591Z * [new branch] gh/kwen2501/231/orig -> origin/gh/kwen2501/231/orig 2025-09-07T07:36:04.0130298Z * [new branch] gh/kwen2501/232/base -> origin/gh/kwen2501/232/base 2025-09-07T07:36:04.0131480Z * [new branch] gh/kwen2501/232/head -> origin/gh/kwen2501/232/head 2025-09-07T07:36:04.0132615Z * [new branch] gh/kwen2501/232/orig -> origin/gh/kwen2501/232/orig 2025-09-07T07:36:04.0134975Z * [new branch] gh/laithsakka/156/base -> origin/gh/laithsakka/156/base 2025-09-07T07:36:04.0136498Z * [new branch] gh/laithsakka/156/head -> origin/gh/laithsakka/156/head 2025-09-07T07:36:04.0137776Z * [new branch] gh/laithsakka/156/orig -> origin/gh/laithsakka/156/orig 2025-09-07T07:36:04.0139588Z * [new branch] gh/laithsakka/160/base -> origin/gh/laithsakka/160/base 2025-09-07T07:36:04.0140775Z * [new branch] gh/laithsakka/160/head -> origin/gh/laithsakka/160/head 2025-09-07T07:36:04.0141945Z * [new branch] gh/laithsakka/160/orig -> origin/gh/laithsakka/160/orig 2025-09-07T07:36:04.0143639Z * [new branch] gh/laithsakka/178/base -> origin/gh/laithsakka/178/base 2025-09-07T07:36:04.0144850Z * [new branch] gh/laithsakka/178/head -> origin/gh/laithsakka/178/head 2025-09-07T07:36:04.0146040Z * [new branch] gh/laithsakka/178/orig -> origin/gh/laithsakka/178/orig 2025-09-07T07:36:04.0147861Z * [new branch] gh/laithsakka/191/base -> origin/gh/laithsakka/191/base 2025-09-07T07:36:04.0149088Z * [new branch] gh/laithsakka/191/head -> origin/gh/laithsakka/191/head 2025-09-07T07:36:04.0150258Z * [new branch] gh/laithsakka/191/orig -> origin/gh/laithsakka/191/orig 2025-09-07T07:36:04.0152054Z * [new branch] gh/laithsakka/237/base -> origin/gh/laithsakka/237/base 2025-09-07T07:36:04.0153292Z * [new branch] gh/laithsakka/237/head -> origin/gh/laithsakka/237/head 2025-09-07T07:36:04.0154466Z * [new branch] gh/laithsakka/237/orig -> origin/gh/laithsakka/237/orig 2025-09-07T07:36:04.0156171Z * [new branch] gh/laithsakka/249/base -> origin/gh/laithsakka/249/base 2025-09-07T07:36:04.0157788Z * [new branch] gh/laithsakka/249/head -> origin/gh/laithsakka/249/head 2025-09-07T07:36:04.0158993Z * [new branch] gh/laithsakka/249/orig -> origin/gh/laithsakka/249/orig 2025-09-07T07:36:04.0160717Z * [new branch] gh/laithsakka/251/base -> origin/gh/laithsakka/251/base 2025-09-07T07:36:04.0161910Z * [new branch] gh/laithsakka/251/head -> origin/gh/laithsakka/251/head 2025-09-07T07:36:04.0163100Z * [new branch] gh/laithsakka/251/orig -> origin/gh/laithsakka/251/orig 2025-09-07T07:36:04.0164858Z * [new branch] gh/laithsakka/254/base -> origin/gh/laithsakka/254/base 2025-09-07T07:36:04.0165976Z * [new branch] gh/laithsakka/254/head -> origin/gh/laithsakka/254/head 2025-09-07T07:36:04.0167174Z * [new branch] gh/laithsakka/254/orig -> origin/gh/laithsakka/254/orig 2025-09-07T07:36:04.0169015Z * [new branch] gh/laithsakka/255/base -> origin/gh/laithsakka/255/base 2025-09-07T07:36:04.0170178Z * [new branch] gh/laithsakka/255/head -> origin/gh/laithsakka/255/head 2025-09-07T07:36:04.0171316Z * [new branch] gh/laithsakka/255/orig -> origin/gh/laithsakka/255/orig 2025-09-07T07:36:04.0173005Z * [new branch] gh/laithsakka/256/base -> origin/gh/laithsakka/256/base 2025-09-07T07:36:04.0174244Z * [new branch] gh/laithsakka/256/head -> origin/gh/laithsakka/256/head 2025-09-07T07:36:04.0175317Z * [new branch] gh/laithsakka/256/orig -> origin/gh/laithsakka/256/orig 2025-09-07T07:36:04.0176998Z * [new branch] gh/laithsakka/257/base -> origin/gh/laithsakka/257/base 2025-09-07T07:36:04.0178251Z * [new branch] gh/laithsakka/257/head -> origin/gh/laithsakka/257/head 2025-09-07T07:36:04.0179454Z * [new branch] gh/laithsakka/257/orig -> origin/gh/laithsakka/257/orig 2025-09-07T07:36:04.0181140Z * [new branch] gh/laithsakka/258/base -> origin/gh/laithsakka/258/base 2025-09-07T07:36:04.0182555Z * [new branch] gh/laithsakka/258/head -> origin/gh/laithsakka/258/head 2025-09-07T07:36:04.0183654Z * [new branch] gh/laithsakka/258/orig -> origin/gh/laithsakka/258/orig 2025-09-07T07:36:04.0185361Z * [new branch] gh/laithsakka/259/base -> origin/gh/laithsakka/259/base 2025-09-07T07:36:04.0186585Z * [new branch] gh/laithsakka/259/head -> origin/gh/laithsakka/259/head 2025-09-07T07:36:04.0187776Z * [new branch] gh/laithsakka/259/orig -> origin/gh/laithsakka/259/orig 2025-09-07T07:36:04.0189384Z * [new branch] gh/laithsakka/260/base -> origin/gh/laithsakka/260/base 2025-09-07T07:36:04.0190564Z * [new branch] gh/laithsakka/260/head -> origin/gh/laithsakka/260/head 2025-09-07T07:36:04.0191907Z * [new branch] gh/laithsakka/260/orig -> origin/gh/laithsakka/260/orig 2025-09-07T07:36:04.0193575Z * [new branch] gh/laithsakka/261/base -> origin/gh/laithsakka/261/base 2025-09-07T07:36:04.0194762Z * [new branch] gh/laithsakka/261/head -> origin/gh/laithsakka/261/head 2025-09-07T07:36:04.0195899Z * [new branch] gh/laithsakka/261/orig -> origin/gh/laithsakka/261/orig 2025-09-07T07:36:04.0197873Z * [new branch] gh/laithsakka/262/base -> origin/gh/laithsakka/262/base 2025-09-07T07:36:04.0199352Z * [new branch] gh/laithsakka/262/head -> origin/gh/laithsakka/262/head 2025-09-07T07:36:04.0200603Z * [new branch] gh/laithsakka/262/orig -> origin/gh/laithsakka/262/orig 2025-09-07T07:36:04.0202688Z * [new branch] gh/laithsakka/263/base -> origin/gh/laithsakka/263/base 2025-09-07T07:36:04.0204187Z * [new branch] gh/laithsakka/263/head -> origin/gh/laithsakka/263/head 2025-09-07T07:36:04.0205398Z * [new branch] gh/laithsakka/263/orig -> origin/gh/laithsakka/263/orig 2025-09-07T07:36:04.0207081Z * [new branch] gh/laithsakka/264/base -> origin/gh/laithsakka/264/base 2025-09-07T07:36:04.0208254Z * [new branch] gh/laithsakka/264/head -> origin/gh/laithsakka/264/head 2025-09-07T07:36:04.0209410Z * [new branch] gh/laithsakka/264/orig -> origin/gh/laithsakka/264/orig 2025-09-07T07:36:04.0211209Z * [new branch] gh/laithsakka/265/base -> origin/gh/laithsakka/265/base 2025-09-07T07:36:04.0212442Z * [new branch] gh/laithsakka/265/head -> origin/gh/laithsakka/265/head 2025-09-07T07:36:04.0213633Z * [new branch] gh/laithsakka/265/orig -> origin/gh/laithsakka/265/orig 2025-09-07T07:36:04.0215369Z * [new branch] gh/laithsakka/266/base -> origin/gh/laithsakka/266/base 2025-09-07T07:36:04.0216576Z * [new branch] gh/laithsakka/266/head -> origin/gh/laithsakka/266/head 2025-09-07T07:36:04.0218016Z * [new branch] gh/laithsakka/266/orig -> origin/gh/laithsakka/266/orig 2025-09-07T07:36:04.0219780Z * [new branch] gh/laithsakka/267/base -> origin/gh/laithsakka/267/base 2025-09-07T07:36:04.0221434Z * [new branch] gh/laithsakka/267/head -> origin/gh/laithsakka/267/head 2025-09-07T07:36:04.0222617Z * [new branch] gh/laithsakka/267/orig -> origin/gh/laithsakka/267/orig 2025-09-07T07:36:04.0224402Z * [new branch] gh/laithsakka/268/base -> origin/gh/laithsakka/268/base 2025-09-07T07:36:04.0225546Z * [new branch] gh/laithsakka/268/head -> origin/gh/laithsakka/268/head 2025-09-07T07:36:04.0226776Z * [new branch] gh/laithsakka/268/orig -> origin/gh/laithsakka/268/orig 2025-09-07T07:36:04.0228600Z * [new branch] gh/laithsakka/28/base -> origin/gh/laithsakka/28/base 2025-09-07T07:36:04.0230205Z * [new branch] gh/laithsakka/29/base -> origin/gh/laithsakka/29/base 2025-09-07T07:36:04.0232028Z * [new branch] gh/laithsakka/30/base -> origin/gh/laithsakka/30/base 2025-09-07T07:36:04.0233157Z * [new branch] gh/laithsakka/30/head -> origin/gh/laithsakka/30/head 2025-09-07T07:36:04.0234786Z * [new branch] gh/laithsakka/31/base -> origin/gh/laithsakka/31/base 2025-09-07T07:36:04.0235917Z * [new branch] gh/laithsakka/31/head -> origin/gh/laithsakka/31/head 2025-09-07T07:36:04.0237572Z * [new branch] gh/laithsakka/32/base -> origin/gh/laithsakka/32/base 2025-09-07T07:36:04.0238700Z * [new branch] gh/laithsakka/32/head -> origin/gh/laithsakka/32/head 2025-09-07T07:36:04.0242174Z * [new branch] gh/lucaskabela/1/base -> origin/gh/lucaskabela/1/base 2025-09-07T07:36:04.0243354Z * [new branch] gh/lucaskabela/1/head -> origin/gh/lucaskabela/1/head 2025-09-07T07:36:04.0245160Z * [new branch] gh/lucaskabela/10/base -> origin/gh/lucaskabela/10/base 2025-09-07T07:36:04.0246370Z * [new branch] gh/lucaskabela/10/head -> origin/gh/lucaskabela/10/head 2025-09-07T07:36:04.0247549Z * [new branch] gh/lucaskabela/10/orig -> origin/gh/lucaskabela/10/orig 2025-09-07T07:36:04.0249114Z * [new branch] gh/lucaskabela/11/base -> origin/gh/lucaskabela/11/base 2025-09-07T07:36:04.0250308Z * [new branch] gh/lucaskabela/11/head -> origin/gh/lucaskabela/11/head 2025-09-07T07:36:04.0251490Z * [new branch] gh/lucaskabela/11/orig -> origin/gh/lucaskabela/11/orig 2025-09-07T07:36:04.0253168Z * [new branch] gh/lucaskabela/12/base -> origin/gh/lucaskabela/12/base 2025-09-07T07:36:04.0254332Z * [new branch] gh/lucaskabela/12/head -> origin/gh/lucaskabela/12/head 2025-09-07T07:36:04.0255642Z * [new branch] gh/lucaskabela/12/orig -> origin/gh/lucaskabela/12/orig 2025-09-07T07:36:04.0257235Z * [new branch] gh/lucaskabela/13/base -> origin/gh/lucaskabela/13/base 2025-09-07T07:36:04.0258533Z * [new branch] gh/lucaskabela/13/head -> origin/gh/lucaskabela/13/head 2025-09-07T07:36:04.0259710Z * [new branch] gh/lucaskabela/13/orig -> origin/gh/lucaskabela/13/orig 2025-09-07T07:36:04.0261347Z * [new branch] gh/lucaskabela/14/base -> origin/gh/lucaskabela/14/base 2025-09-07T07:36:04.0263018Z * [new branch] gh/lucaskabela/14/head -> origin/gh/lucaskabela/14/head 2025-09-07T07:36:04.0264242Z * [new branch] gh/lucaskabela/14/orig -> origin/gh/lucaskabela/14/orig 2025-09-07T07:36:04.0265853Z * [new branch] gh/lucaskabela/15/base -> origin/gh/lucaskabela/15/base 2025-09-07T07:36:04.0267027Z * [new branch] gh/lucaskabela/15/head -> origin/gh/lucaskabela/15/head 2025-09-07T07:36:04.0268252Z * [new branch] gh/lucaskabela/15/orig -> origin/gh/lucaskabela/15/orig 2025-09-07T07:36:04.0269995Z * [new branch] gh/lucaskabela/16/base -> origin/gh/lucaskabela/16/base 2025-09-07T07:36:04.0271183Z * [new branch] gh/lucaskabela/16/head -> origin/gh/lucaskabela/16/head 2025-09-07T07:36:04.0272364Z * [new branch] gh/lucaskabela/16/orig -> origin/gh/lucaskabela/16/orig 2025-09-07T07:36:04.0273988Z * [new branch] gh/lucaskabela/17/base -> origin/gh/lucaskabela/17/base 2025-09-07T07:36:04.0275096Z * [new branch] gh/lucaskabela/17/head -> origin/gh/lucaskabela/17/head 2025-09-07T07:36:04.0276230Z * [new branch] gh/lucaskabela/17/orig -> origin/gh/lucaskabela/17/orig 2025-09-07T07:36:04.0277922Z * [new branch] gh/lucaskabela/2/base -> origin/gh/lucaskabela/2/base 2025-09-07T07:36:04.0279090Z * [new branch] gh/lucaskabela/2/head -> origin/gh/lucaskabela/2/head 2025-09-07T07:36:04.0280281Z * [new branch] gh/lucaskabela/2/orig -> origin/gh/lucaskabela/2/orig 2025-09-07T07:36:04.0282066Z * [new branch] gh/lucaskabela/3/base -> origin/gh/lucaskabela/3/base 2025-09-07T07:36:04.0283196Z * [new branch] gh/lucaskabela/3/head -> origin/gh/lucaskabela/3/head 2025-09-07T07:36:04.0284352Z * [new branch] gh/lucaskabela/3/orig -> origin/gh/lucaskabela/3/orig 2025-09-07T07:36:04.0286038Z * [new branch] gh/lucaskabela/4/base -> origin/gh/lucaskabela/4/base 2025-09-07T07:36:04.0287208Z * [new branch] gh/lucaskabela/4/head -> origin/gh/lucaskabela/4/head 2025-09-07T07:36:04.0288417Z * [new branch] gh/lucaskabela/4/orig -> origin/gh/lucaskabela/4/orig 2025-09-07T07:36:04.0290264Z * [new branch] gh/lucaskabela/5/base -> origin/gh/lucaskabela/5/base 2025-09-07T07:36:04.0291433Z * [new branch] gh/lucaskabela/5/head -> origin/gh/lucaskabela/5/head 2025-09-07T07:36:04.0292612Z * [new branch] gh/lucaskabela/5/orig -> origin/gh/lucaskabela/5/orig 2025-09-07T07:36:04.0294241Z * [new branch] gh/lucaskabela/6/base -> origin/gh/lucaskabela/6/base 2025-09-07T07:36:04.0295462Z * [new branch] gh/lucaskabela/6/head -> origin/gh/lucaskabela/6/head 2025-09-07T07:36:04.0296656Z * [new branch] gh/lucaskabela/6/orig -> origin/gh/lucaskabela/6/orig 2025-09-07T07:36:04.0298525Z * [new branch] gh/lucaskabela/7/base -> origin/gh/lucaskabela/7/base 2025-09-07T07:36:04.0299675Z * [new branch] gh/lucaskabela/7/head -> origin/gh/lucaskabela/7/head 2025-09-07T07:36:04.0300865Z * [new branch] gh/lucaskabela/7/orig -> origin/gh/lucaskabela/7/orig 2025-09-07T07:36:04.0302554Z * [new branch] gh/lucaskabela/8/base -> origin/gh/lucaskabela/8/base 2025-09-07T07:36:04.0304011Z * [new branch] gh/lucaskabela/8/head -> origin/gh/lucaskabela/8/head 2025-09-07T07:36:04.0305219Z * [new branch] gh/lucaskabela/8/orig -> origin/gh/lucaskabela/8/orig 2025-09-07T07:36:04.0306913Z * [new branch] gh/lucaskabela/9/base -> origin/gh/lucaskabela/9/base 2025-09-07T07:36:04.0308273Z * [new branch] gh/lucaskabela/9/head -> origin/gh/lucaskabela/9/head 2025-09-07T07:36:04.0309234Z * [new branch] gh/lucaskabela/9/orig -> origin/gh/lucaskabela/9/orig 2025-09-07T07:36:04.0311404Z * [new branch] gh/lw/3/base -> origin/gh/lw/3/base 2025-09-07T07:36:04.0312600Z * [new branch] gh/lw/3/head -> origin/gh/lw/3/head 2025-09-07T07:36:04.0313988Z * [new branch] gh/lw/3/orig -> origin/gh/lw/3/orig 2025-09-07T07:36:04.0315910Z * [new branch] gh/malfet/14/base -> origin/gh/malfet/14/base 2025-09-07T07:36:04.0318189Z * [new branch] gh/malfet/330/base -> origin/gh/malfet/330/base 2025-09-07T07:36:04.0319413Z * [new branch] gh/malfet/330/head -> origin/gh/malfet/330/head 2025-09-07T07:36:04.0320602Z * [new branch] gh/malfet/330/orig -> origin/gh/malfet/330/orig 2025-09-07T07:36:04.0322273Z * [new branch] gh/malfet/396/base -> origin/gh/malfet/396/base 2025-09-07T07:36:04.0323464Z * [new branch] gh/malfet/396/head -> origin/gh/malfet/396/head 2025-09-07T07:36:04.0324660Z * [new branch] gh/malfet/396/orig -> origin/gh/malfet/396/orig 2025-09-07T07:36:04.0326475Z * [new branch] gh/malfet/397/base -> origin/gh/malfet/397/base 2025-09-07T07:36:04.0327705Z * [new branch] gh/malfet/397/head -> origin/gh/malfet/397/head 2025-09-07T07:36:04.0328865Z * [new branch] gh/malfet/397/orig -> origin/gh/malfet/397/orig 2025-09-07T07:36:04.0330571Z * [new branch] gh/malfet/398/base -> origin/gh/malfet/398/base 2025-09-07T07:36:04.0331761Z * [new branch] gh/malfet/398/head -> origin/gh/malfet/398/head 2025-09-07T07:36:04.0332866Z * [new branch] gh/malfet/398/orig -> origin/gh/malfet/398/orig 2025-09-07T07:36:04.0334605Z * [new branch] gh/malfet/399/base -> origin/gh/malfet/399/base 2025-09-07T07:36:04.0337945Z * [new branch] gh/malfet/399/head -> origin/gh/malfet/399/head 2025-09-07T07:36:04.0338390Z * [new branch] gh/malfet/399/orig -> origin/gh/malfet/399/orig 2025-09-07T07:36:04.0340631Z * [new branch] gh/malfet/414/base -> origin/gh/malfet/414/base 2025-09-07T07:36:04.0340767Z * [new branch] gh/malfet/414/head -> origin/gh/malfet/414/head 2025-09-07T07:36:04.0341612Z * [new branch] gh/malfet/414/orig -> origin/gh/malfet/414/orig 2025-09-07T07:36:04.0343263Z * [new branch] gh/malfet/417/base -> origin/gh/malfet/417/base 2025-09-07T07:36:04.0344440Z * [new branch] gh/malfet/417/head -> origin/gh/malfet/417/head 2025-09-07T07:36:04.0345647Z * [new branch] gh/malfet/417/orig -> origin/gh/malfet/417/orig 2025-09-07T07:36:04.0347243Z * [new branch] gh/malfet/418/base -> origin/gh/malfet/418/base 2025-09-07T07:36:04.0348393Z * [new branch] gh/malfet/418/head -> origin/gh/malfet/418/head 2025-09-07T07:36:04.0349585Z * [new branch] gh/malfet/418/orig -> origin/gh/malfet/418/orig 2025-09-07T07:36:04.0351384Z * [new branch] gh/malfet/475/base -> origin/gh/malfet/475/base 2025-09-07T07:36:04.0353196Z * [new branch] gh/malfet/475/head -> origin/gh/malfet/475/head 2025-09-07T07:36:04.0354402Z * [new branch] gh/malfet/475/orig -> origin/gh/malfet/475/orig 2025-09-07T07:36:04.0356486Z * [new branch] gh/malfet/476/base -> origin/gh/malfet/476/base 2025-09-07T07:36:04.0357686Z * [new branch] gh/malfet/476/head -> origin/gh/malfet/476/head 2025-09-07T07:36:04.0358894Z * [new branch] gh/malfet/476/orig -> origin/gh/malfet/476/orig 2025-09-07T07:36:04.0360508Z * [new branch] gh/malfet/477/base -> origin/gh/malfet/477/base 2025-09-07T07:36:04.0361724Z * [new branch] gh/malfet/477/head -> origin/gh/malfet/477/head 2025-09-07T07:36:04.0363194Z * [new branch] gh/malfet/477/orig -> origin/gh/malfet/477/orig 2025-09-07T07:36:04.0364776Z * [new branch] gh/malfet/478/base -> origin/gh/malfet/478/base 2025-09-07T07:36:04.0365941Z * [new branch] gh/malfet/478/head -> origin/gh/malfet/478/head 2025-09-07T07:36:04.0367133Z * [new branch] gh/malfet/478/orig -> origin/gh/malfet/478/orig 2025-09-07T07:36:04.0368785Z * [new branch] gh/malfet/479/base -> origin/gh/malfet/479/base 2025-09-07T07:36:04.0370012Z * [new branch] gh/malfet/479/head -> origin/gh/malfet/479/head 2025-09-07T07:36:04.0371215Z * [new branch] gh/malfet/479/orig -> origin/gh/malfet/479/orig 2025-09-07T07:36:04.0373325Z * [new branch] gh/malfet/480/base -> origin/gh/malfet/480/base 2025-09-07T07:36:04.0374548Z * [new branch] gh/malfet/480/head -> origin/gh/malfet/480/head 2025-09-07T07:36:04.0375746Z * [new branch] gh/malfet/480/orig -> origin/gh/malfet/480/orig 2025-09-07T07:36:04.0377496Z * [new branch] gh/malfet/481/base -> origin/gh/malfet/481/base 2025-09-07T07:36:04.0378779Z * [new branch] gh/malfet/481/head -> origin/gh/malfet/481/head 2025-09-07T07:36:04.0379929Z * [new branch] gh/malfet/481/orig -> origin/gh/malfet/481/orig 2025-09-07T07:36:04.0381542Z * [new branch] gh/malfet/482/base -> origin/gh/malfet/482/base 2025-09-07T07:36:04.0383202Z * [new branch] gh/malfet/482/head -> origin/gh/malfet/482/head 2025-09-07T07:36:04.0384360Z * [new branch] gh/malfet/482/orig -> origin/gh/malfet/482/orig 2025-09-07T07:36:04.0386500Z * [new branch] gh/malfet/483/base -> origin/gh/malfet/483/base 2025-09-07T07:36:04.0387700Z * [new branch] gh/malfet/483/head -> origin/gh/malfet/483/head 2025-09-07T07:36:04.0388930Z * [new branch] gh/malfet/483/orig -> origin/gh/malfet/483/orig 2025-09-07T07:36:04.0390695Z * [new branch] gh/malfet/484/base -> origin/gh/malfet/484/base 2025-09-07T07:36:04.0391865Z * [new branch] gh/malfet/484/head -> origin/gh/malfet/484/head 2025-09-07T07:36:04.0393093Z * [new branch] gh/malfet/484/orig -> origin/gh/malfet/484/orig 2025-09-07T07:36:04.0394919Z * [new branch] gh/malfet/485/base -> origin/gh/malfet/485/base 2025-09-07T07:36:04.0396108Z * [new branch] gh/malfet/485/head -> origin/gh/malfet/485/head 2025-09-07T07:36:04.0397328Z * [new branch] gh/malfet/485/orig -> origin/gh/malfet/485/orig 2025-09-07T07:36:04.0399221Z * [new branch] gh/malfet/486/base -> origin/gh/malfet/486/base 2025-09-07T07:36:04.0400402Z * [new branch] gh/malfet/486/head -> origin/gh/malfet/486/head 2025-09-07T07:36:04.0401581Z * [new branch] gh/malfet/486/orig -> origin/gh/malfet/486/orig 2025-09-07T07:36:04.0403512Z * [new branch] gh/malfet/487/base -> origin/gh/malfet/487/base 2025-09-07T07:36:04.0408574Z * [new branch] gh/malfet/487/head -> origin/gh/malfet/487/head 2025-09-07T07:36:04.0409032Z * [new branch] gh/malfet/487/orig -> origin/gh/malfet/487/orig 2025-09-07T07:36:04.0411650Z * [new branch] gh/malfet/488/base -> origin/gh/malfet/488/base 2025-09-07T07:36:04.0412898Z * [new branch] gh/malfet/488/head -> origin/gh/malfet/488/head 2025-09-07T07:36:04.0414158Z * [new branch] gh/malfet/488/orig -> origin/gh/malfet/488/orig 2025-09-07T07:36:04.0416088Z * [new branch] gh/malfet/489/base -> origin/gh/malfet/489/base 2025-09-07T07:36:04.0417497Z * [new branch] gh/malfet/489/head -> origin/gh/malfet/489/head 2025-09-07T07:36:04.0418863Z * [new branch] gh/malfet/489/orig -> origin/gh/malfet/489/orig 2025-09-07T07:36:04.0420983Z * [new branch] gh/malfet/490/base -> origin/gh/malfet/490/base 2025-09-07T07:36:04.0422296Z * [new branch] gh/malfet/490/head -> origin/gh/malfet/490/head 2025-09-07T07:36:04.0423717Z * [new branch] gh/malfet/490/orig -> origin/gh/malfet/490/orig 2025-09-07T07:36:04.0425608Z * [new branch] gh/malfet/491/base -> origin/gh/malfet/491/base 2025-09-07T07:36:04.0426846Z * [new branch] gh/malfet/491/head -> origin/gh/malfet/491/head 2025-09-07T07:36:04.0428122Z * [new branch] gh/malfet/491/orig -> origin/gh/malfet/491/orig 2025-09-07T07:36:04.0429843Z * [new branch] gh/malfet/492/base -> origin/gh/malfet/492/base 2025-09-07T07:36:04.0431199Z * [new branch] gh/malfet/492/head -> origin/gh/malfet/492/head 2025-09-07T07:36:04.0432430Z * [new branch] gh/malfet/492/orig -> origin/gh/malfet/492/orig 2025-09-07T07:36:04.0434415Z * [new branch] gh/malfet/493/base -> origin/gh/malfet/493/base 2025-09-07T07:36:04.0435446Z * [new branch] gh/malfet/493/head -> origin/gh/malfet/493/head 2025-09-07T07:36:04.0436682Z * [new branch] gh/malfet/493/orig -> origin/gh/malfet/493/orig 2025-09-07T07:36:04.0438612Z * [new branch] gh/malfet/494/base -> origin/gh/malfet/494/base 2025-09-07T07:36:04.0440011Z * [new branch] gh/malfet/494/head -> origin/gh/malfet/494/head 2025-09-07T07:36:04.0441237Z * [new branch] gh/malfet/494/orig -> origin/gh/malfet/494/orig 2025-09-07T07:36:04.0442980Z * [new branch] gh/malfet/495/base -> origin/gh/malfet/495/base 2025-09-07T07:36:04.0444238Z * [new branch] gh/malfet/495/head -> origin/gh/malfet/495/head 2025-09-07T07:36:04.0445451Z * [new branch] gh/malfet/495/orig -> origin/gh/malfet/495/orig 2025-09-07T07:36:04.0447297Z * [new branch] gh/malfet/496/base -> origin/gh/malfet/496/base 2025-09-07T07:36:04.0448500Z * [new branch] gh/malfet/496/head -> origin/gh/malfet/496/head 2025-09-07T07:36:04.0449764Z * [new branch] gh/malfet/496/orig -> origin/gh/malfet/496/orig 2025-09-07T07:36:04.0451543Z * [new branch] gh/malfet/497/base -> origin/gh/malfet/497/base 2025-09-07T07:36:04.0452749Z * [new branch] gh/malfet/497/head -> origin/gh/malfet/497/head 2025-09-07T07:36:04.0454017Z * [new branch] gh/malfet/497/orig -> origin/gh/malfet/497/orig 2025-09-07T07:36:04.0455800Z * [new branch] gh/malfet/498/base -> origin/gh/malfet/498/base 2025-09-07T07:36:04.0456982Z * [new branch] gh/malfet/498/head -> origin/gh/malfet/498/head 2025-09-07T07:36:04.0458247Z * [new branch] gh/malfet/498/orig -> origin/gh/malfet/498/orig 2025-09-07T07:36:04.0459969Z * [new branch] gh/malfet/499/base -> origin/gh/malfet/499/base 2025-09-07T07:36:04.0461211Z * [new branch] gh/malfet/499/head -> origin/gh/malfet/499/head 2025-09-07T07:36:04.0462542Z * [new branch] gh/malfet/499/orig -> origin/gh/malfet/499/orig 2025-09-07T07:36:04.0464352Z * [new branch] gh/malfet/500/base -> origin/gh/malfet/500/base 2025-09-07T07:36:04.0465546Z * [new branch] gh/malfet/500/head -> origin/gh/malfet/500/head 2025-09-07T07:36:04.0466822Z * [new branch] gh/malfet/500/orig -> origin/gh/malfet/500/orig 2025-09-07T07:36:04.0469190Z * [new branch] gh/malfet/501/base -> origin/gh/malfet/501/base 2025-09-07T07:36:04.0470370Z * [new branch] gh/malfet/501/head -> origin/gh/malfet/501/head 2025-09-07T07:36:04.0471553Z * [new branch] gh/malfet/501/orig -> origin/gh/malfet/501/orig 2025-09-07T07:36:04.0473441Z * [new branch] gh/malfet/502/base -> origin/gh/malfet/502/base 2025-09-07T07:36:04.0474617Z * [new branch] gh/malfet/502/head -> origin/gh/malfet/502/head 2025-09-07T07:36:04.0475850Z * [new branch] gh/malfet/502/orig -> origin/gh/malfet/502/orig 2025-09-07T07:36:04.0477847Z * [new branch] gh/malfet/503/base -> origin/gh/malfet/503/base 2025-09-07T07:36:04.0479078Z * [new branch] gh/malfet/503/head -> origin/gh/malfet/503/head 2025-09-07T07:36:04.0480253Z * [new branch] gh/malfet/503/orig -> origin/gh/malfet/503/orig 2025-09-07T07:36:04.0482059Z * [new branch] gh/malfet/504/base -> origin/gh/malfet/504/base 2025-09-07T07:36:04.0483229Z * [new branch] gh/malfet/504/head -> origin/gh/malfet/504/head 2025-09-07T07:36:04.0484537Z * [new branch] gh/malfet/504/orig -> origin/gh/malfet/504/orig 2025-09-07T07:36:04.0486302Z * [new branch] gh/malfet/505/base -> origin/gh/malfet/505/base 2025-09-07T07:36:04.0487496Z * [new branch] gh/malfet/505/head -> origin/gh/malfet/505/head 2025-09-07T07:36:04.0488684Z * [new branch] gh/malfet/505/orig -> origin/gh/malfet/505/orig 2025-09-07T07:36:04.0490554Z * [new branch] gh/malfet/506/base -> origin/gh/malfet/506/base 2025-09-07T07:36:04.0491678Z * [new branch] gh/malfet/506/head -> origin/gh/malfet/506/head 2025-09-07T07:36:04.0492849Z * [new branch] gh/malfet/506/orig -> origin/gh/malfet/506/orig 2025-09-07T07:36:04.0494702Z * [new branch] gh/malfet/507/base -> origin/gh/malfet/507/base 2025-09-07T07:36:04.0495883Z * [new branch] gh/malfet/507/head -> origin/gh/malfet/507/head 2025-09-07T07:36:04.0497109Z * [new branch] gh/malfet/507/orig -> origin/gh/malfet/507/orig 2025-09-07T07:36:04.0499073Z * [new branch] gh/malfet/508/base -> origin/gh/malfet/508/base 2025-09-07T07:36:04.0500304Z * [new branch] gh/malfet/508/head -> origin/gh/malfet/508/head 2025-09-07T07:36:04.0501525Z * [new branch] gh/malfet/508/orig -> origin/gh/malfet/508/orig 2025-09-07T07:36:04.0503342Z * [new branch] gh/malfet/509/base -> origin/gh/malfet/509/base 2025-09-07T07:36:04.0504775Z * [new branch] gh/malfet/509/head -> origin/gh/malfet/509/head 2025-09-07T07:36:04.0506052Z * [new branch] gh/malfet/509/orig -> origin/gh/malfet/509/orig 2025-09-07T07:36:04.0507902Z * [new branch] gh/malfet/510/base -> origin/gh/malfet/510/base 2025-09-07T07:36:04.0509097Z * [new branch] gh/malfet/510/head -> origin/gh/malfet/510/head 2025-09-07T07:36:04.0510295Z * [new branch] gh/malfet/510/orig -> origin/gh/malfet/510/orig 2025-09-07T07:36:04.0512576Z * [new branch] gh/malfet/511/base -> origin/gh/malfet/511/base 2025-09-07T07:36:04.0514051Z * [new branch] gh/malfet/511/head -> origin/gh/malfet/511/head 2025-09-07T07:36:04.0515244Z * [new branch] gh/malfet/511/orig -> origin/gh/malfet/511/orig 2025-09-07T07:36:04.0517016Z * [new branch] gh/malfet/512/base -> origin/gh/malfet/512/base 2025-09-07T07:36:04.0518243Z * [new branch] gh/malfet/512/head -> origin/gh/malfet/512/head 2025-09-07T07:36:04.0519502Z * [new branch] gh/malfet/512/orig -> origin/gh/malfet/512/orig 2025-09-07T07:36:04.0521317Z * [new branch] gh/malfet/513/base -> origin/gh/malfet/513/base 2025-09-07T07:36:04.0522502Z * [new branch] gh/malfet/513/head -> origin/gh/malfet/513/head 2025-09-07T07:36:04.0523648Z * [new branch] gh/malfet/513/orig -> origin/gh/malfet/513/orig 2025-09-07T07:36:04.0525604Z * [new branch] gh/malfet/64/base -> origin/gh/malfet/64/base 2025-09-07T07:36:04.0526857Z * [new branch] gh/malfet/64/head -> origin/gh/malfet/64/head 2025-09-07T07:36:04.0529353Z * [new branch] gh/manuelcandales/10/base -> origin/gh/manuelcandales/10/base 2025-09-07T07:36:04.0530474Z * [new branch] gh/manuelcandales/10/head -> origin/gh/manuelcandales/10/head 2025-09-07T07:36:04.0531628Z * [new branch] gh/manuelcandales/10/orig -> origin/gh/manuelcandales/10/orig 2025-09-07T07:36:04.0533345Z * [new branch] gh/manuelcandales/11/base -> origin/gh/manuelcandales/11/base 2025-09-07T07:36:04.0534555Z * [new branch] gh/manuelcandales/11/head -> origin/gh/manuelcandales/11/head 2025-09-07T07:36:04.0535805Z * [new branch] gh/manuelcandales/11/orig -> origin/gh/manuelcandales/11/orig 2025-09-07T07:36:04.0537518Z * [new branch] gh/manuelcandales/9/base -> origin/gh/manuelcandales/9/base 2025-09-07T07:36:04.0538786Z * [new branch] gh/manuelcandales/9/head -> origin/gh/manuelcandales/9/head 2025-09-07T07:36:04.0540042Z * [new branch] gh/manuelcandales/9/orig -> origin/gh/manuelcandales/9/orig 2025-09-07T07:36:04.0542477Z * [new branch] gh/markkm/1/base -> origin/gh/markkm/1/base 2025-09-07T07:36:04.0544480Z * [new branch] gh/masnesral/204/base -> origin/gh/masnesral/204/base 2025-09-07T07:36:04.0546205Z * [new branch] gh/masnesral/204/head -> origin/gh/masnesral/204/head 2025-09-07T07:36:04.0547129Z * [new branch] gh/masnesral/204/orig -> origin/gh/masnesral/204/orig 2025-09-07T07:36:04.0548957Z * [new branch] gh/masnesral/235/base -> origin/gh/masnesral/235/base 2025-09-07T07:36:04.0550169Z * [new branch] gh/masnesral/235/head -> origin/gh/masnesral/235/head 2025-09-07T07:36:04.0551591Z * [new branch] gh/masnesral/235/orig -> origin/gh/masnesral/235/orig 2025-09-07T07:36:04.0553362Z * [new branch] gh/masnesral/34/base -> origin/gh/masnesral/34/base 2025-09-07T07:36:04.0555540Z * [new branch] gh/mhorowitz/0/base -> origin/gh/mhorowitz/0/base 2025-09-07T07:36:04.0556725Z * [new branch] gh/mhorowitz/0/head -> origin/gh/mhorowitz/0/head 2025-09-07T07:36:04.0558320Z * [new branch] gh/mhorowitz/1/base -> origin/gh/mhorowitz/1/base 2025-09-07T07:36:04.0559490Z * [new branch] gh/mhorowitz/1/head -> origin/gh/mhorowitz/1/head 2025-09-07T07:36:04.0561635Z * [new branch] gh/mhorowitz/2/base -> origin/gh/mhorowitz/2/base 2025-09-07T07:36:04.0562312Z * [new branch] gh/mhorowitz/2/head -> origin/gh/mhorowitz/2/head 2025-09-07T07:36:04.0563948Z * [new branch] gh/mhorowitz/3/base -> origin/gh/mhorowitz/3/base 2025-09-07T07:36:04.0565152Z * [new branch] gh/mhorowitz/3/head -> origin/gh/mhorowitz/3/head 2025-09-07T07:36:04.0566793Z * [new branch] gh/mhorowitz/4/base -> origin/gh/mhorowitz/4/base 2025-09-07T07:36:04.0567977Z * [new branch] gh/mhorowitz/4/head -> origin/gh/mhorowitz/4/head 2025-09-07T07:36:04.0569558Z * [new branch] gh/mhorowitz/5/base -> origin/gh/mhorowitz/5/base 2025-09-07T07:36:04.0570687Z * [new branch] gh/mhorowitz/5/head -> origin/gh/mhorowitz/5/head 2025-09-07T07:36:04.0572249Z * [new branch] gh/mhorowitz/6/base -> origin/gh/mhorowitz/6/base 2025-09-07T07:36:04.0573408Z * [new branch] gh/mhorowitz/6/head -> origin/gh/mhorowitz/6/head 2025-09-07T07:36:04.0575585Z * [new branch] gh/mikaylagawarecki/234/base -> origin/gh/mikaylagawarecki/234/base 2025-09-07T07:36:04.0576784Z * [new branch] gh/mikaylagawarecki/234/head -> origin/gh/mikaylagawarecki/234/head 2025-09-07T07:36:04.0578490Z * [new branch] gh/mikaylagawarecki/235/base -> origin/gh/mikaylagawarecki/235/base 2025-09-07T07:36:04.0579709Z * [new branch] gh/mikaylagawarecki/235/head -> origin/gh/mikaylagawarecki/235/head 2025-09-07T07:36:04.0581424Z * [new branch] gh/mikaylagawarecki/236/base -> origin/gh/mikaylagawarecki/236/base 2025-09-07T07:36:04.0582568Z * [new branch] gh/mikaylagawarecki/236/head -> origin/gh/mikaylagawarecki/236/head 2025-09-07T07:36:04.0584253Z * [new branch] gh/mikaylagawarecki/237/base -> origin/gh/mikaylagawarecki/237/base 2025-09-07T07:36:04.0585365Z * [new branch] gh/mikaylagawarecki/237/head -> origin/gh/mikaylagawarecki/237/head 2025-09-07T07:36:04.0587017Z * [new branch] gh/mikaylagawarecki/238/base -> origin/gh/mikaylagawarecki/238/base 2025-09-07T07:36:04.0588393Z * [new branch] gh/mikaylagawarecki/238/head -> origin/gh/mikaylagawarecki/238/head 2025-09-07T07:36:04.0590197Z * [new branch] gh/mikaylagawarecki/317/base -> origin/gh/mikaylagawarecki/317/base 2025-09-07T07:36:04.0591428Z * [new branch] gh/mikaylagawarecki/317/head -> origin/gh/mikaylagawarecki/317/head 2025-09-07T07:36:04.0592672Z * [new branch] gh/mikaylagawarecki/317/orig -> origin/gh/mikaylagawarecki/317/orig 2025-09-07T07:36:04.0594379Z * [new branch] gh/mikaylagawarecki/320/base -> origin/gh/mikaylagawarecki/320/base 2025-09-07T07:36:04.0595561Z * [new branch] gh/mikaylagawarecki/320/head -> origin/gh/mikaylagawarecki/320/head 2025-09-07T07:36:04.0596806Z * [new branch] gh/mikaylagawarecki/320/orig -> origin/gh/mikaylagawarecki/320/orig 2025-09-07T07:36:04.0598524Z * [new branch] gh/mikaylagawarecki/329/base -> origin/gh/mikaylagawarecki/329/base 2025-09-07T07:36:04.0599712Z * [new branch] gh/mikaylagawarecki/329/head -> origin/gh/mikaylagawarecki/329/head 2025-09-07T07:36:04.0600885Z * [new branch] gh/mikaylagawarecki/329/orig -> origin/gh/mikaylagawarecki/329/orig 2025-09-07T07:36:04.0602584Z * [new branch] gh/mikaylagawarecki/330/base -> origin/gh/mikaylagawarecki/330/base 2025-09-07T07:36:04.0603996Z * [new branch] gh/mikaylagawarecki/330/head -> origin/gh/mikaylagawarecki/330/head 2025-09-07T07:36:04.0605407Z * [new branch] gh/mikaylagawarecki/330/orig -> origin/gh/mikaylagawarecki/330/orig 2025-09-07T07:36:04.0607135Z * [new branch] gh/mikaylagawarecki/331/base -> origin/gh/mikaylagawarecki/331/base 2025-09-07T07:36:04.0608339Z * [new branch] gh/mikaylagawarecki/331/head -> origin/gh/mikaylagawarecki/331/head 2025-09-07T07:36:04.0609506Z * [new branch] gh/mikaylagawarecki/331/orig -> origin/gh/mikaylagawarecki/331/orig 2025-09-07T07:36:04.0611415Z * [new branch] gh/mikaylagawarecki/332/base -> origin/gh/mikaylagawarecki/332/base 2025-09-07T07:36:04.0612573Z * [new branch] gh/mikaylagawarecki/332/head -> origin/gh/mikaylagawarecki/332/head 2025-09-07T07:36:04.0613848Z * [new branch] gh/mikaylagawarecki/332/orig -> origin/gh/mikaylagawarecki/332/orig 2025-09-07T07:36:04.0616481Z * [new branch] gh/mikaylagawarecki/334/base -> origin/gh/mikaylagawarecki/334/base 2025-09-07T07:36:04.0617637Z * [new branch] gh/mikaylagawarecki/334/head -> origin/gh/mikaylagawarecki/334/head 2025-09-07T07:36:04.0618861Z * [new branch] gh/mikaylagawarecki/334/orig -> origin/gh/mikaylagawarecki/334/orig 2025-09-07T07:36:04.0620596Z * [new branch] gh/mikaylagawarecki/335/base -> origin/gh/mikaylagawarecki/335/base 2025-09-07T07:36:04.0621840Z * [new branch] gh/mikaylagawarecki/335/head -> origin/gh/mikaylagawarecki/335/head 2025-09-07T07:36:04.0623012Z * [new branch] gh/mikaylagawarecki/335/orig -> origin/gh/mikaylagawarecki/335/orig 2025-09-07T07:36:04.0624938Z * [new branch] gh/mikaylagawarecki/336/base -> origin/gh/mikaylagawarecki/336/base 2025-09-07T07:36:04.0626183Z * [new branch] gh/mikaylagawarecki/336/head -> origin/gh/mikaylagawarecki/336/head 2025-09-07T07:36:04.0627351Z * [new branch] gh/mikaylagawarecki/336/orig -> origin/gh/mikaylagawarecki/336/orig 2025-09-07T07:36:04.0628953Z * [new branch] gh/mikaylagawarecki/337/base -> origin/gh/mikaylagawarecki/337/base 2025-09-07T07:36:04.0630417Z * [new branch] gh/mikaylagawarecki/337/head -> origin/gh/mikaylagawarecki/337/head 2025-09-07T07:36:04.0631703Z * [new branch] gh/mikaylagawarecki/337/orig -> origin/gh/mikaylagawarecki/337/orig 2025-09-07T07:36:04.0633617Z * [new branch] gh/mikaylagawarecki/338/base -> origin/gh/mikaylagawarecki/338/base 2025-09-07T07:36:04.0634847Z * [new branch] gh/mikaylagawarecki/338/head -> origin/gh/mikaylagawarecki/338/head 2025-09-07T07:36:04.0636038Z * [new branch] gh/mikaylagawarecki/338/orig -> origin/gh/mikaylagawarecki/338/orig 2025-09-07T07:36:04.0638161Z * [new branch] gh/mikaylagawarecki/339/base -> origin/gh/mikaylagawarecki/339/base 2025-09-07T07:36:04.0639369Z * [new branch] gh/mikaylagawarecki/339/head -> origin/gh/mikaylagawarecki/339/head 2025-09-07T07:36:04.0640957Z * [new branch] gh/mikaylagawarecki/339/orig -> origin/gh/mikaylagawarecki/339/orig 2025-09-07T07:36:04.0642761Z * [new branch] gh/mlazos/1/base -> origin/gh/mlazos/1/base 2025-09-07T07:36:04.0643991Z * [new branch] gh/mlazos/1/head -> origin/gh/mlazos/1/head 2025-09-07T07:36:04.0645175Z * [new branch] gh/mlazos/1/orig -> origin/gh/mlazos/1/orig 2025-09-07T07:36:04.0646937Z * [new branch] gh/mlazos/12/base -> origin/gh/mlazos/12/base 2025-09-07T07:36:04.0648229Z * [new branch] gh/mlazos/12/head -> origin/gh/mlazos/12/head 2025-09-07T07:36:04.0649438Z * [new branch] gh/mlazos/12/orig -> origin/gh/mlazos/12/orig 2025-09-07T07:36:04.0651250Z * [new branch] gh/mlazos/13/base -> origin/gh/mlazos/13/base 2025-09-07T07:36:04.0652467Z * [new branch] gh/mlazos/13/head -> origin/gh/mlazos/13/head 2025-09-07T07:36:04.0653673Z * [new branch] gh/mlazos/13/orig -> origin/gh/mlazos/13/orig 2025-09-07T07:36:04.0655386Z * [new branch] gh/mlazos/14/base -> origin/gh/mlazos/14/base 2025-09-07T07:36:04.0656580Z * [new branch] gh/mlazos/14/head -> origin/gh/mlazos/14/head 2025-09-07T07:36:04.0657868Z * [new branch] gh/mlazos/14/orig -> origin/gh/mlazos/14/orig 2025-09-07T07:36:04.0659656Z * [new branch] gh/mlazos/15/base -> origin/gh/mlazos/15/base 2025-09-07T07:36:04.0661017Z * [new branch] gh/mlazos/15/head -> origin/gh/mlazos/15/head 2025-09-07T07:36:04.0662187Z * [new branch] gh/mlazos/15/orig -> origin/gh/mlazos/15/orig 2025-09-07T07:36:04.0663987Z * [new branch] gh/mlazos/16/base -> origin/gh/mlazos/16/base 2025-09-07T07:36:04.0665293Z * [new branch] gh/mlazos/16/head -> origin/gh/mlazos/16/head 2025-09-07T07:36:04.0666464Z * [new branch] gh/mlazos/16/orig -> origin/gh/mlazos/16/orig 2025-09-07T07:36:04.0668114Z * [new branch] gh/mlazos/17/base -> origin/gh/mlazos/17/base 2025-09-07T07:36:04.0669245Z * [new branch] gh/mlazos/17/head -> origin/gh/mlazos/17/head 2025-09-07T07:36:04.0670440Z * [new branch] gh/mlazos/17/orig -> origin/gh/mlazos/17/orig 2025-09-07T07:36:04.0672306Z * [new branch] gh/mlazos/2/base -> origin/gh/mlazos/2/base 2025-09-07T07:36:04.0673408Z * [new branch] gh/mlazos/2/head -> origin/gh/mlazos/2/head 2025-09-07T07:36:04.0674609Z * [new branch] gh/mlazos/2/orig -> origin/gh/mlazos/2/orig 2025-09-07T07:36:04.0676946Z * [new branch] gh/mlazos/3/base -> origin/gh/mlazos/3/base 2025-09-07T07:36:04.0678090Z * [new branch] gh/mlazos/3/head -> origin/gh/mlazos/3/head 2025-09-07T07:36:04.0679260Z * [new branch] gh/mlazos/3/orig -> origin/gh/mlazos/3/orig 2025-09-07T07:36:04.0681503Z * [new branch] gh/mrmiywj/1/base -> origin/gh/mrmiywj/1/base 2025-09-07T07:36:04.0682864Z * [new branch] gh/mrmiywj/1/head -> origin/gh/mrmiywj/1/head 2025-09-07T07:36:04.0685052Z * [new branch] gh/muchulee8/62/base -> origin/gh/muchulee8/62/base 2025-09-07T07:36:04.0686311Z * [new branch] gh/muchulee8/62/head -> origin/gh/muchulee8/62/head 2025-09-07T07:36:04.0687512Z * [new branch] gh/muchulee8/62/orig -> origin/gh/muchulee8/62/orig 2025-09-07T07:36:04.0689247Z * [new branch] gh/muchulee8/63/base -> origin/gh/muchulee8/63/base 2025-09-07T07:36:04.0690417Z * [new branch] gh/muchulee8/63/head -> origin/gh/muchulee8/63/head 2025-09-07T07:36:04.0691686Z * [new branch] gh/muchulee8/63/orig -> origin/gh/muchulee8/63/orig 2025-09-07T07:36:04.0693463Z * [new branch] gh/muchulee8/64/base -> origin/gh/muchulee8/64/base 2025-09-07T07:36:04.0694615Z * [new branch] gh/muchulee8/64/head -> origin/gh/muchulee8/64/head 2025-09-07T07:36:04.0695861Z * [new branch] gh/muchulee8/64/orig -> origin/gh/muchulee8/64/orig 2025-09-07T07:36:04.0697751Z * [new branch] gh/muchulee8/65/base -> origin/gh/muchulee8/65/base 2025-09-07T07:36:04.0699198Z * [new branch] gh/muchulee8/65/head -> origin/gh/muchulee8/65/head 2025-09-07T07:36:04.0700898Z * [new branch] gh/muchulee8/65/orig -> origin/gh/muchulee8/65/orig 2025-09-07T07:36:04.0703241Z * [new branch] gh/naveenthangudu/1/base -> origin/gh/naveenthangudu/1/base 2025-09-07T07:36:04.0704529Z * [new branch] gh/naveenthangudu/1/head -> origin/gh/naveenthangudu/1/head 2025-09-07T07:36:04.0705802Z * [new branch] gh/naveenthangudu/1/orig -> origin/gh/naveenthangudu/1/orig 2025-09-07T07:36:04.0707483Z * [new branch] gh/naveenthangudu/2/base -> origin/gh/naveenthangudu/2/base 2025-09-07T07:36:04.0708663Z * [new branch] gh/naveenthangudu/2/head -> origin/gh/naveenthangudu/2/head 2025-09-07T07:36:04.0709893Z * [new branch] gh/naveenthangudu/2/orig -> origin/gh/naveenthangudu/2/orig 2025-09-07T07:36:04.0711560Z * [new branch] gh/naveenthangudu/3/base -> origin/gh/naveenthangudu/3/base 2025-09-07T07:36:04.0712835Z * [new branch] gh/naveenthangudu/3/head -> origin/gh/naveenthangudu/3/head 2025-09-07T07:36:04.0714045Z * [new branch] gh/naveenthangudu/3/orig -> origin/gh/naveenthangudu/3/orig 2025-09-07T07:36:04.0715742Z * [new branch] gh/naveenthangudu/4/base -> origin/gh/naveenthangudu/4/base 2025-09-07T07:36:04.0716962Z * [new branch] gh/naveenthangudu/4/head -> origin/gh/naveenthangudu/4/head 2025-09-07T07:36:04.0718202Z * [new branch] gh/naveenthangudu/4/orig -> origin/gh/naveenthangudu/4/orig 2025-09-07T07:36:04.0719905Z * [new branch] gh/naveenthangudu/5/base -> origin/gh/naveenthangudu/5/base 2025-09-07T07:36:04.0721061Z * [new branch] gh/naveenthangudu/5/head -> origin/gh/naveenthangudu/5/head 2025-09-07T07:36:04.0722326Z * [new branch] gh/naveenthangudu/5/orig -> origin/gh/naveenthangudu/5/orig 2025-09-07T07:36:04.0723987Z * [new branch] gh/naveenthangudu/6/base -> origin/gh/naveenthangudu/6/base 2025-09-07T07:36:04.0725178Z * [new branch] gh/naveenthangudu/6/head -> origin/gh/naveenthangudu/6/head 2025-09-07T07:36:04.0726311Z * [new branch] gh/naveenthangudu/6/orig -> origin/gh/naveenthangudu/6/orig 2025-09-07T07:36:04.0728401Z * [new branch] gh/oulgen/35/base -> origin/gh/oulgen/35/base 2025-09-07T07:36:04.0729633Z * [new branch] gh/oulgen/35/head -> origin/gh/oulgen/35/head 2025-09-07T07:36:04.0730845Z * [new branch] gh/oulgen/35/orig -> origin/gh/oulgen/35/orig 2025-09-07T07:36:04.0732546Z * [new branch] gh/oulgen/48/base -> origin/gh/oulgen/48/base 2025-09-07T07:36:04.0733700Z * [new branch] gh/oulgen/48/head -> origin/gh/oulgen/48/head 2025-09-07T07:36:04.0735088Z * [new branch] gh/oulgen/48/orig -> origin/gh/oulgen/48/orig 2025-09-07T07:36:04.0737200Z * [new branch] gh/oulgen/49/base -> origin/gh/oulgen/49/base 2025-09-07T07:36:04.0738439Z * [new branch] gh/oulgen/49/head -> origin/gh/oulgen/49/head 2025-09-07T07:36:04.0739649Z * [new branch] gh/oulgen/49/orig -> origin/gh/oulgen/49/orig 2025-09-07T07:36:04.0741963Z * [new branch] gh/pearu/108/base -> origin/gh/pearu/108/base 2025-09-07T07:36:04.0743278Z * [new branch] gh/pearu/108/head -> origin/gh/pearu/108/head 2025-09-07T07:36:04.0744556Z * [new branch] gh/pearu/108/orig -> origin/gh/pearu/108/orig 2025-09-07T07:36:04.0746671Z * [new branch] gh/pearu/109/base -> origin/gh/pearu/109/base 2025-09-07T07:36:04.0747924Z * [new branch] gh/pearu/109/head -> origin/gh/pearu/109/head 2025-09-07T07:36:04.0749106Z * [new branch] gh/pearu/109/orig -> origin/gh/pearu/109/orig 2025-09-07T07:36:04.0750836Z * [new branch] gh/pearu/110/base -> origin/gh/pearu/110/base 2025-09-07T07:36:04.0752044Z * [new branch] gh/pearu/110/head -> origin/gh/pearu/110/head 2025-09-07T07:36:04.0753189Z * [new branch] gh/pearu/110/orig -> origin/gh/pearu/110/orig 2025-09-07T07:36:04.0755081Z * [new branch] gh/pearu/111/base -> origin/gh/pearu/111/base 2025-09-07T07:36:04.0756254Z * [new branch] gh/pearu/111/head -> origin/gh/pearu/111/head 2025-09-07T07:36:04.0757464Z * [new branch] gh/pearu/111/orig -> origin/gh/pearu/111/orig 2025-09-07T07:36:04.0759076Z * [new branch] gh/pearu/112/base -> origin/gh/pearu/112/base 2025-09-07T07:36:04.0760270Z * [new branch] gh/pearu/112/head -> origin/gh/pearu/112/head 2025-09-07T07:36:04.0761525Z * [new branch] gh/pearu/112/orig -> origin/gh/pearu/112/orig 2025-09-07T07:36:04.0763169Z * [new branch] gh/pearu/113/base -> origin/gh/pearu/113/base 2025-09-07T07:36:04.0764448Z * [new branch] gh/pearu/113/head -> origin/gh/pearu/113/head 2025-09-07T07:36:04.0765612Z * [new branch] gh/pearu/113/orig -> origin/gh/pearu/113/orig 2025-09-07T07:36:04.0767325Z * [new branch] gh/pearu/114/base -> origin/gh/pearu/114/base 2025-09-07T07:36:04.0768474Z * [new branch] gh/pearu/114/head -> origin/gh/pearu/114/head 2025-09-07T07:36:04.0769729Z * [new branch] gh/pearu/114/orig -> origin/gh/pearu/114/orig 2025-09-07T07:36:04.0771648Z * [new branch] gh/pearu/115/base -> origin/gh/pearu/115/base 2025-09-07T07:36:04.0772845Z * [new branch] gh/pearu/115/head -> origin/gh/pearu/115/head 2025-09-07T07:36:04.0774021Z * [new branch] gh/pearu/115/orig -> origin/gh/pearu/115/orig 2025-09-07T07:36:04.0775741Z * [new branch] gh/pearu/116/base -> origin/gh/pearu/116/base 2025-09-07T07:36:04.0776877Z * [new branch] gh/pearu/116/head -> origin/gh/pearu/116/head 2025-09-07T07:36:04.0778146Z * [new branch] gh/pearu/116/orig -> origin/gh/pearu/116/orig 2025-09-07T07:36:04.0779851Z * [new branch] gh/pearu/117/base -> origin/gh/pearu/117/base 2025-09-07T07:36:04.0781030Z * [new branch] gh/pearu/117/head -> origin/gh/pearu/117/head 2025-09-07T07:36:04.0782149Z * [new branch] gh/pearu/117/orig -> origin/gh/pearu/117/orig 2025-09-07T07:36:04.0784194Z * [new branch] gh/pearu/56/base -> origin/gh/pearu/56/base 2025-09-07T07:36:04.0785517Z * [new branch] gh/pearu/56/head -> origin/gh/pearu/56/head 2025-09-07T07:36:04.0786744Z * [new branch] gh/pearu/56/orig -> origin/gh/pearu/56/orig 2025-09-07T07:36:04.0788601Z * [new branch] gh/pearu/97/base -> origin/gh/pearu/97/base 2025-09-07T07:36:04.0789806Z * [new branch] gh/pearu/97/head -> origin/gh/pearu/97/head 2025-09-07T07:36:04.0791073Z * [new branch] gh/pearu/97/orig -> origin/gh/pearu/97/orig 2025-09-07T07:36:04.0793237Z * [new branch] gh/qqaatw/29/base -> origin/gh/qqaatw/29/base 2025-09-07T07:36:04.0794954Z * [new branch] gh/qqaatw/29/head -> origin/gh/qqaatw/29/head 2025-09-07T07:36:04.0796114Z * [new branch] gh/qqaatw/29/orig -> origin/gh/qqaatw/29/orig 2025-09-07T07:36:04.0797921Z * [new branch] gh/raymo/refresh-script -> origin/gh/raymo/refresh-script 2025-09-07T07:36:04.0800067Z * [new branch] gh/rec/141/base -> origin/gh/rec/141/base 2025-09-07T07:36:04.0801288Z * [new branch] gh/rec/141/head -> origin/gh/rec/141/head 2025-09-07T07:36:04.0802982Z * [new branch] gh/rec/153/base -> origin/gh/rec/153/base 2025-09-07T07:36:04.0804405Z * [new branch] gh/rec/153/head -> origin/gh/rec/153/head 2025-09-07T07:36:04.0805598Z * [new branch] gh/rec/153/orig -> origin/gh/rec/153/orig 2025-09-07T07:36:04.0807279Z * [new branch] gh/rec/154/base -> origin/gh/rec/154/base 2025-09-07T07:36:04.0808610Z * [new branch] gh/rec/154/head -> origin/gh/rec/154/head 2025-09-07T07:36:04.0809790Z * [new branch] gh/rec/154/orig -> origin/gh/rec/154/orig 2025-09-07T07:36:04.0811444Z * [new branch] gh/rec/156/base -> origin/gh/rec/156/base 2025-09-07T07:36:04.0813094Z * [new branch] gh/rec/156/head -> origin/gh/rec/156/head 2025-09-07T07:36:04.0814381Z * [new branch] gh/rec/156/orig -> origin/gh/rec/156/orig 2025-09-07T07:36:04.0816095Z * [new branch] gh/rec/160/base -> origin/gh/rec/160/base 2025-09-07T07:36:04.0817272Z * [new branch] gh/rec/160/head -> origin/gh/rec/160/head 2025-09-07T07:36:04.0818516Z * [new branch] gh/rec/160/orig -> origin/gh/rec/160/orig 2025-09-07T07:36:04.0820237Z * [new branch] gh/rec/162/base -> origin/gh/rec/162/base 2025-09-07T07:36:04.0821424Z * [new branch] gh/rec/162/head -> origin/gh/rec/162/head 2025-09-07T07:36:04.0822614Z * [new branch] gh/rec/162/orig -> origin/gh/rec/162/orig 2025-09-07T07:36:04.0824268Z * [new branch] gh/rec/163/base -> origin/gh/rec/163/base 2025-09-07T07:36:04.0825432Z * [new branch] gh/rec/163/head -> origin/gh/rec/163/head 2025-09-07T07:36:04.0826588Z * [new branch] gh/rec/163/orig -> origin/gh/rec/163/orig 2025-09-07T07:36:04.0828255Z * [new branch] gh/rec/164/base -> origin/gh/rec/164/base 2025-09-07T07:36:04.0829507Z * [new branch] gh/rec/164/head -> origin/gh/rec/164/head 2025-09-07T07:36:04.0830712Z * [new branch] gh/rec/164/orig -> origin/gh/rec/164/orig 2025-09-07T07:36:04.0832444Z * [new branch] gh/rec/165/base -> origin/gh/rec/165/base 2025-09-07T07:36:04.0833739Z * [new branch] gh/rec/165/head -> origin/gh/rec/165/head 2025-09-07T07:36:04.0834899Z * [new branch] gh/rec/165/orig -> origin/gh/rec/165/orig 2025-09-07T07:36:04.0836554Z * [new branch] gh/rec/166/base -> origin/gh/rec/166/base 2025-09-07T07:36:04.0837768Z * [new branch] gh/rec/166/head -> origin/gh/rec/166/head 2025-09-07T07:36:04.0838927Z * [new branch] gh/rec/166/orig -> origin/gh/rec/166/orig 2025-09-07T07:36:04.0841105Z * [new branch] gh/robert-hardwick/1/base -> origin/gh/robert-hardwick/1/base 2025-09-07T07:36:04.0842311Z * [new branch] gh/robert-hardwick/1/head -> origin/gh/robert-hardwick/1/head 2025-09-07T07:36:04.0843684Z * [new branch] gh/robert-hardwick/1/orig -> origin/gh/robert-hardwick/1/orig 2025-09-07T07:36:04.0845429Z * [new branch] gh/robert-hardwick/2/base -> origin/gh/robert-hardwick/2/base 2025-09-07T07:36:04.0846835Z * [new branch] gh/robert-hardwick/2/head -> origin/gh/robert-hardwick/2/head 2025-09-07T07:36:04.0847907Z * [new branch] gh/robert-hardwick/2/orig -> origin/gh/robert-hardwick/2/orig 2025-09-07T07:36:04.0849574Z * [new branch] gh/robert-hardwick/3/base -> origin/gh/robert-hardwick/3/base 2025-09-07T07:36:04.0850766Z * [new branch] gh/robert-hardwick/3/head -> origin/gh/robert-hardwick/3/head 2025-09-07T07:36:04.0852086Z * [new branch] gh/robert-hardwick/3/orig -> origin/gh/robert-hardwick/3/orig 2025-09-07T07:36:04.0853704Z * [new branch] gh/robert-hardwick/4/base -> origin/gh/robert-hardwick/4/base 2025-09-07T07:36:04.0854861Z * [new branch] gh/robert-hardwick/4/head -> origin/gh/robert-hardwick/4/head 2025-09-07T07:36:04.0856052Z * [new branch] gh/robert-hardwick/4/orig -> origin/gh/robert-hardwick/4/orig 2025-09-07T07:36:04.0858235Z * [new branch] gh/rtimpe/1/base -> origin/gh/rtimpe/1/base 2025-09-07T07:36:04.0859482Z * [new branch] gh/rtimpe/1/head -> origin/gh/rtimpe/1/head 2025-09-07T07:36:04.0861291Z * [new branch] gh/rtimpe/10/base -> origin/gh/rtimpe/10/base 2025-09-07T07:36:04.0862466Z * [new branch] gh/rtimpe/10/head -> origin/gh/rtimpe/10/head 2025-09-07T07:36:04.0863761Z * [new branch] gh/rtimpe/10/orig -> origin/gh/rtimpe/10/orig 2025-09-07T07:36:04.0865477Z * [new branch] gh/rtimpe/11/base -> origin/gh/rtimpe/11/base 2025-09-07T07:36:04.0866672Z * [new branch] gh/rtimpe/11/head -> origin/gh/rtimpe/11/head 2025-09-07T07:36:04.0867833Z * [new branch] gh/rtimpe/11/orig -> origin/gh/rtimpe/11/orig 2025-09-07T07:36:04.0869499Z * [new branch] gh/rtimpe/12/base -> origin/gh/rtimpe/12/base 2025-09-07T07:36:04.0870696Z * [new branch] gh/rtimpe/12/head -> origin/gh/rtimpe/12/head 2025-09-07T07:36:04.0872325Z * [new branch] gh/rtimpe/12/orig -> origin/gh/rtimpe/12/orig 2025-09-07T07:36:04.0873988Z * [new branch] gh/rtimpe/13/base -> origin/gh/rtimpe/13/base 2025-09-07T07:36:04.0875264Z * [new branch] gh/rtimpe/13/head -> origin/gh/rtimpe/13/head 2025-09-07T07:36:04.0876426Z * [new branch] gh/rtimpe/13/orig -> origin/gh/rtimpe/13/orig 2025-09-07T07:36:04.0878084Z * [new branch] gh/rtimpe/14/base -> origin/gh/rtimpe/14/base 2025-09-07T07:36:04.0879488Z * [new branch] gh/rtimpe/14/head -> origin/gh/rtimpe/14/head 2025-09-07T07:36:04.0880707Z * [new branch] gh/rtimpe/14/orig -> origin/gh/rtimpe/14/orig 2025-09-07T07:36:04.0882382Z * [new branch] gh/rtimpe/15/base -> origin/gh/rtimpe/15/base 2025-09-07T07:36:04.0883556Z * [new branch] gh/rtimpe/15/head -> origin/gh/rtimpe/15/head 2025-09-07T07:36:04.0884757Z * [new branch] gh/rtimpe/15/orig -> origin/gh/rtimpe/15/orig 2025-09-07T07:36:04.0886386Z * [new branch] gh/rtimpe/2/base -> origin/gh/rtimpe/2/base 2025-09-07T07:36:04.0887535Z * [new branch] gh/rtimpe/2/head -> origin/gh/rtimpe/2/head 2025-09-07T07:36:04.0889134Z * [new branch] gh/rtimpe/3/base -> origin/gh/rtimpe/3/base 2025-09-07T07:36:04.0890289Z * [new branch] gh/rtimpe/3/head -> origin/gh/rtimpe/3/head 2025-09-07T07:36:04.0891971Z * [new branch] gh/rtimpe/4/base -> origin/gh/rtimpe/4/base 2025-09-07T07:36:04.0893139Z * [new branch] gh/rtimpe/4/head -> origin/gh/rtimpe/4/head 2025-09-07T07:36:04.0894929Z * [new branch] gh/rtimpe/9/base -> origin/gh/rtimpe/9/base 2025-09-07T07:36:04.0896130Z * [new branch] gh/rtimpe/9/head -> origin/gh/rtimpe/9/head 2025-09-07T07:36:04.0897394Z * [new branch] gh/rtimpe/9/orig -> origin/gh/rtimpe/9/orig 2025-09-07T07:36:04.0899552Z * [new branch] gh/ruisizhang123/1/base -> origin/gh/ruisizhang123/1/base 2025-09-07T07:36:04.0900750Z * [new branch] gh/ruisizhang123/1/head -> origin/gh/ruisizhang123/1/head 2025-09-07T07:36:04.0901935Z * [new branch] gh/ruisizhang123/1/orig -> origin/gh/ruisizhang123/1/orig 2025-09-07T07:36:04.0903832Z * [new branch] gh/ruisizhang123/4/base -> origin/gh/ruisizhang123/4/base 2025-09-07T07:36:04.0905023Z * [new branch] gh/ruisizhang123/4/head -> origin/gh/ruisizhang123/4/head 2025-09-07T07:36:04.0906188Z * [new branch] gh/ruisizhang123/4/orig -> origin/gh/ruisizhang123/4/orig 2025-09-07T07:36:04.0908414Z * [new branch] gh/ruisizhang123/5/base -> origin/gh/ruisizhang123/5/base 2025-09-07T07:36:04.0909740Z * [new branch] gh/ruisizhang123/5/head -> origin/gh/ruisizhang123/5/head 2025-09-07T07:36:04.0910901Z * [new branch] gh/ruisizhang123/5/orig -> origin/gh/ruisizhang123/5/orig 2025-09-07T07:36:04.0912738Z * [new branch] gh/ruisizhang123/6/base -> origin/gh/ruisizhang123/6/base 2025-09-07T07:36:04.0913723Z * [new branch] gh/ruisizhang123/6/head -> origin/gh/ruisizhang123/6/head 2025-09-07T07:36:04.0914976Z * [new branch] gh/ruisizhang123/6/orig -> origin/gh/ruisizhang123/6/orig 2025-09-07T07:36:04.0916958Z * [new branch] gh/ruisizhang123/7/base -> origin/gh/ruisizhang123/7/base 2025-09-07T07:36:04.0930155Z * [new branch] gh/ruisizhang123/7/head -> origin/gh/ruisizhang123/7/head 2025-09-07T07:36:04.0930555Z * [new branch] gh/ruisizhang123/7/orig -> origin/gh/ruisizhang123/7/orig 2025-09-07T07:36:04.0931064Z * [new branch] gh/ruisizhang123/8/base -> origin/gh/ruisizhang123/8/base 2025-09-07T07:36:04.0931406Z * [new branch] gh/ruisizhang123/8/head -> origin/gh/ruisizhang123/8/head 2025-09-07T07:36:04.0931742Z * [new branch] gh/ruisizhang123/8/orig -> origin/gh/ruisizhang123/8/orig 2025-09-07T07:36:04.0932073Z * [new branch] gh/ruisizhang123/9/base -> origin/gh/ruisizhang123/9/base 2025-09-07T07:36:04.0932397Z * [new branch] gh/ruisizhang123/9/head -> origin/gh/ruisizhang123/9/head 2025-09-07T07:36:04.0932725Z * [new branch] gh/ruisizhang123/9/orig -> origin/gh/ruisizhang123/9/orig 2025-09-07T07:36:04.0933050Z * [new branch] gh/sarckk/2/base -> origin/gh/sarckk/2/base 2025-09-07T07:36:04.0933362Z * [new branch] gh/sarckk/2/head -> origin/gh/sarckk/2/head 2025-09-07T07:36:04.0933662Z * [new branch] gh/sarckk/2/orig -> origin/gh/sarckk/2/orig 2025-09-07T07:36:04.0934449Z * [new branch] gh/seemethere/35/base -> origin/gh/seemethere/35/base 2025-09-07T07:36:04.0935690Z * [new branch] gh/seemethere/35/head -> origin/gh/seemethere/35/head 2025-09-07T07:36:04.0936924Z * [new branch] gh/seemethere/35/orig -> origin/gh/seemethere/35/orig 2025-09-07T07:36:04.0938704Z * [new branch] gh/seemethere/37/base -> origin/gh/seemethere/37/base 2025-09-07T07:36:04.0939891Z * [new branch] gh/seemethere/37/head -> origin/gh/seemethere/37/head 2025-09-07T07:36:04.0941054Z * [new branch] gh/seemethere/37/orig -> origin/gh/seemethere/37/orig 2025-09-07T07:36:04.0942718Z * [new branch] gh/seemethere/43/base -> origin/gh/seemethere/43/base 2025-09-07T07:36:04.0943996Z * [new branch] gh/seemethere/43/head -> origin/gh/seemethere/43/head 2025-09-07T07:36:04.0945186Z * [new branch] gh/seemethere/43/orig -> origin/gh/seemethere/43/orig 2025-09-07T07:36:04.0946906Z * [new branch] gh/seemethere/44/base -> origin/gh/seemethere/44/base 2025-09-07T07:36:04.0948020Z * [new branch] gh/seemethere/44/head -> origin/gh/seemethere/44/head 2025-09-07T07:36:04.0949222Z * [new branch] gh/seemethere/44/orig -> origin/gh/seemethere/44/orig 2025-09-07T07:36:04.0950914Z * [new branch] gh/seemethere/48/base -> origin/gh/seemethere/48/base 2025-09-07T07:36:04.0952298Z * [new branch] gh/seemethere/48/head -> origin/gh/seemethere/48/head 2025-09-07T07:36:04.0953465Z * [new branch] gh/seemethere/48/orig -> origin/gh/seemethere/48/orig 2025-09-07T07:36:04.0955130Z * [new branch] gh/seemethere/49/base -> origin/gh/seemethere/49/base 2025-09-07T07:36:04.0956311Z * [new branch] gh/seemethere/49/head -> origin/gh/seemethere/49/head 2025-09-07T07:36:04.0957582Z * [new branch] gh/seemethere/49/orig -> origin/gh/seemethere/49/orig 2025-09-07T07:36:04.0959276Z * [new branch] gh/seemethere/52/base -> origin/gh/seemethere/52/base 2025-09-07T07:36:04.0960581Z * [new branch] gh/seemethere/52/head -> origin/gh/seemethere/52/head 2025-09-07T07:36:04.0961762Z * [new branch] gh/seemethere/52/orig -> origin/gh/seemethere/52/orig 2025-09-07T07:36:04.0963456Z * [new branch] gh/seemethere/53/base -> origin/gh/seemethere/53/base 2025-09-07T07:36:04.0964632Z * [new branch] gh/seemethere/53/head -> origin/gh/seemethere/53/head 2025-09-07T07:36:04.0965814Z * [new branch] gh/seemethere/53/orig -> origin/gh/seemethere/53/orig 2025-09-07T07:36:04.0967491Z * [new branch] gh/seemethere/54/base -> origin/gh/seemethere/54/base 2025-09-07T07:36:04.0968694Z * [new branch] gh/seemethere/54/head -> origin/gh/seemethere/54/head 2025-09-07T07:36:04.0969896Z * [new branch] gh/seemethere/54/orig -> origin/gh/seemethere/54/orig 2025-09-07T07:36:04.0971506Z * [new branch] gh/seemethere/55/base -> origin/gh/seemethere/55/base 2025-09-07T07:36:04.0972605Z * [new branch] gh/seemethere/55/head -> origin/gh/seemethere/55/head 2025-09-07T07:36:04.0973834Z * [new branch] gh/seemethere/55/orig -> origin/gh/seemethere/55/orig 2025-09-07T07:36:04.0975511Z * [new branch] gh/seemethere/56/base -> origin/gh/seemethere/56/base 2025-09-07T07:36:04.0976787Z * [new branch] gh/seemethere/56/head -> origin/gh/seemethere/56/head 2025-09-07T07:36:04.0978070Z * [new branch] gh/seemethere/56/orig -> origin/gh/seemethere/56/orig 2025-09-07T07:36:04.0979782Z * [new branch] gh/seemethere/57/base -> origin/gh/seemethere/57/base 2025-09-07T07:36:04.0980939Z * [new branch] gh/seemethere/57/head -> origin/gh/seemethere/57/head 2025-09-07T07:36:04.0982093Z * [new branch] gh/seemethere/57/orig -> origin/gh/seemethere/57/orig 2025-09-07T07:36:04.0983753Z * [new branch] gh/seemethere/58/base -> origin/gh/seemethere/58/base 2025-09-07T07:36:04.0984948Z * [new branch] gh/seemethere/58/head -> origin/gh/seemethere/58/head 2025-09-07T07:36:04.0986288Z * [new branch] gh/seemethere/58/orig -> origin/gh/seemethere/58/orig 2025-09-07T07:36:04.0987965Z * [new branch] gh/seemethere/59/base -> origin/gh/seemethere/59/base 2025-09-07T07:36:04.0989143Z * [new branch] gh/seemethere/59/head -> origin/gh/seemethere/59/head 2025-09-07T07:36:04.0990333Z * [new branch] gh/seemethere/59/orig -> origin/gh/seemethere/59/orig 2025-09-07T07:36:04.0992015Z * [new branch] gh/seemethere/60/base -> origin/gh/seemethere/60/base 2025-09-07T07:36:04.0993243Z * [new branch] gh/seemethere/60/head -> origin/gh/seemethere/60/head 2025-09-07T07:36:04.0994512Z * [new branch] gh/seemethere/60/orig -> origin/gh/seemethere/60/orig 2025-09-07T07:36:04.0996120Z * [new branch] gh/seemethere/61/base -> origin/gh/seemethere/61/base 2025-09-07T07:36:04.0997327Z * [new branch] gh/seemethere/61/head -> origin/gh/seemethere/61/head 2025-09-07T07:36:04.0998612Z * [new branch] gh/seemethere/61/orig -> origin/gh/seemethere/61/orig 2025-09-07T07:36:04.1000268Z * [new branch] gh/seemethere/62/base -> origin/gh/seemethere/62/base 2025-09-07T07:36:04.1001462Z * [new branch] gh/seemethere/62/head -> origin/gh/seemethere/62/head 2025-09-07T07:36:04.1002599Z * [new branch] gh/seemethere/62/orig -> origin/gh/seemethere/62/orig 2025-09-07T07:36:04.1004501Z * [new branch] gh/seemethere/63/base -> origin/gh/seemethere/63/base 2025-09-07T07:36:04.1005666Z * [new branch] gh/seemethere/63/head -> origin/gh/seemethere/63/head 2025-09-07T07:36:04.1006880Z * [new branch] gh/seemethere/63/orig -> origin/gh/seemethere/63/orig 2025-09-07T07:36:04.1009322Z * [new branch] gh/shunting314/145/base -> origin/gh/shunting314/145/base 2025-09-07T07:36:04.1010571Z * [new branch] gh/shunting314/145/head -> origin/gh/shunting314/145/head 2025-09-07T07:36:04.1011824Z * [new branch] gh/shunting314/145/orig -> origin/gh/shunting314/145/orig 2025-09-07T07:36:04.1013745Z * [new branch] gh/shunting314/176/base -> origin/gh/shunting314/176/base 2025-09-07T07:36:04.1015024Z * [new branch] gh/shunting314/176/head -> origin/gh/shunting314/176/head 2025-09-07T07:36:04.1016305Z * [new branch] gh/shunting314/176/orig -> origin/gh/shunting314/176/orig 2025-09-07T07:36:04.1018143Z * [new branch] gh/shunting314/211/base -> origin/gh/shunting314/211/base 2025-09-07T07:36:04.1019361Z * [new branch] gh/shunting314/211/head -> origin/gh/shunting314/211/head 2025-09-07T07:36:04.1020559Z * [new branch] gh/shunting314/211/orig -> origin/gh/shunting314/211/orig 2025-09-07T07:36:04.1022389Z * [new branch] gh/shunting314/212/base -> origin/gh/shunting314/212/base 2025-09-07T07:36:04.1023594Z * [new branch] gh/shunting314/212/head -> origin/gh/shunting314/212/head 2025-09-07T07:36:04.1024797Z * [new branch] gh/shunting314/212/orig -> origin/gh/shunting314/212/orig 2025-09-07T07:36:04.1026967Z * [new branch] gh/shunting314/213/base -> origin/gh/shunting314/213/base 2025-09-07T07:36:04.1028191Z * [new branch] gh/shunting314/213/head -> origin/gh/shunting314/213/head 2025-09-07T07:36:04.1029422Z * [new branch] gh/shunting314/213/orig -> origin/gh/shunting314/213/orig 2025-09-07T07:36:04.1031125Z * [new branch] gh/shunting314/214/base -> origin/gh/shunting314/214/base 2025-09-07T07:36:04.1032346Z * [new branch] gh/shunting314/214/head -> origin/gh/shunting314/214/head 2025-09-07T07:36:04.1033521Z * [new branch] gh/shunting314/214/orig -> origin/gh/shunting314/214/orig 2025-09-07T07:36:04.1035363Z * [new branch] gh/shunting314/215/base -> origin/gh/shunting314/215/base 2025-09-07T07:36:04.1036628Z * [new branch] gh/shunting314/215/head -> origin/gh/shunting314/215/head 2025-09-07T07:36:04.1037821Z * [new branch] gh/shunting314/215/orig -> origin/gh/shunting314/215/orig 2025-09-07T07:36:04.1039481Z * [new branch] gh/shunting314/216/base -> origin/gh/shunting314/216/base 2025-09-07T07:36:04.1040644Z * [new branch] gh/shunting314/216/head -> origin/gh/shunting314/216/head 2025-09-07T07:36:04.1041799Z * [new branch] gh/shunting314/216/orig -> origin/gh/shunting314/216/orig 2025-09-07T07:36:04.1044182Z * [new branch] gh/shunting314/217/base -> origin/gh/shunting314/217/base 2025-09-07T07:36:04.1045294Z * [new branch] gh/shunting314/217/head -> origin/gh/shunting314/217/head 2025-09-07T07:36:04.1046484Z * [new branch] gh/shunting314/217/orig -> origin/gh/shunting314/217/orig 2025-09-07T07:36:04.1048260Z * [new branch] gh/shunting314/218/base -> origin/gh/shunting314/218/base 2025-09-07T07:36:04.1049432Z * [new branch] gh/shunting314/218/head -> origin/gh/shunting314/218/head 2025-09-07T07:36:04.1050608Z * [new branch] gh/shunting314/218/orig -> origin/gh/shunting314/218/orig 2025-09-07T07:36:04.1052206Z * [new branch] gh/shunting314/219/base -> origin/gh/shunting314/219/base 2025-09-07T07:36:04.1053378Z * [new branch] gh/shunting314/219/head -> origin/gh/shunting314/219/head 2025-09-07T07:36:04.1054548Z * [new branch] gh/shunting314/219/orig -> origin/gh/shunting314/219/orig 2025-09-07T07:36:04.1056359Z * [new branch] gh/shunting314/220/base -> origin/gh/shunting314/220/base 2025-09-07T07:36:04.1057945Z * [new branch] gh/shunting314/220/head -> origin/gh/shunting314/220/head 2025-09-07T07:36:04.1059172Z * [new branch] gh/shunting314/220/orig -> origin/gh/shunting314/220/orig 2025-09-07T07:36:04.1060875Z * [new branch] gh/shunting314/221/base -> origin/gh/shunting314/221/base 2025-09-07T07:36:04.1062444Z * [new branch] gh/shunting314/221/head -> origin/gh/shunting314/221/head 2025-09-07T07:36:04.1063680Z * [new branch] gh/shunting314/221/orig -> origin/gh/shunting314/221/orig 2025-09-07T07:36:04.1065703Z * [new branch] gh/shunting314/222/base -> origin/gh/shunting314/222/base 2025-09-07T07:36:04.1066870Z * [new branch] gh/shunting314/222/head -> origin/gh/shunting314/222/head 2025-09-07T07:36:04.1068061Z * [new branch] gh/shunting314/222/orig -> origin/gh/shunting314/222/orig 2025-09-07T07:36:04.1069976Z * [new branch] gh/shunting314/223/base -> origin/gh/shunting314/223/base 2025-09-07T07:36:04.1071242Z * [new branch] gh/shunting314/223/head -> origin/gh/shunting314/223/head 2025-09-07T07:36:04.1072431Z * [new branch] gh/shunting314/223/orig -> origin/gh/shunting314/223/orig 2025-09-07T07:36:04.1074607Z * [new branch] gh/silverguo/1/base -> origin/gh/silverguo/1/base 2025-09-07T07:36:04.1075896Z * [new branch] gh/silverguo/1/head -> origin/gh/silverguo/1/head 2025-09-07T07:36:04.1077902Z * [new branch] gh/silverguo/2/base -> origin/gh/silverguo/2/base 2025-09-07T07:36:04.1079034Z * [new branch] gh/silverguo/2/head -> origin/gh/silverguo/2/head 2025-09-07T07:36:04.1080622Z * [new branch] gh/silverguo/3/base -> origin/gh/silverguo/3/base 2025-09-07T07:36:04.1081808Z * [new branch] gh/silverguo/3/head -> origin/gh/silverguo/3/head 2025-09-07T07:36:04.1083379Z * [new branch] gh/silverguo/4/base -> origin/gh/silverguo/4/base 2025-09-07T07:36:04.1084549Z * [new branch] gh/silverguo/4/head -> origin/gh/silverguo/4/head 2025-09-07T07:36:04.1086735Z * [new branch] gh/sinhaanhsul/1/base -> origin/gh/sinhaanhsul/1/base 2025-09-07T07:36:04.1087921Z * [new branch] gh/sinhaanhsul/1/head -> origin/gh/sinhaanhsul/1/head 2025-09-07T07:36:04.1090049Z * [new branch] gh/skarjala/17/base -> origin/gh/skarjala/17/base 2025-09-07T07:36:04.1091224Z * [new branch] gh/skarjala/17/head -> origin/gh/skarjala/17/head 2025-09-07T07:36:04.1092554Z * [new branch] gh/skarjala/17/orig -> origin/gh/skarjala/17/orig 2025-09-07T07:36:04.1094764Z * [new branch] gh/skarjala/18/base -> origin/gh/skarjala/18/base 2025-09-07T07:36:04.1096151Z * [new branch] gh/skarjala/18/head -> origin/gh/skarjala/18/head 2025-09-07T07:36:04.1097256Z * [new branch] gh/skarjala/18/orig -> origin/gh/skarjala/18/orig 2025-09-07T07:36:04.1099059Z * [new branch] gh/skarjala/19/base -> origin/gh/skarjala/19/base 2025-09-07T07:36:04.1100256Z * [new branch] gh/skarjala/19/head -> origin/gh/skarjala/19/head 2025-09-07T07:36:04.1101469Z * [new branch] gh/skarjala/19/orig -> origin/gh/skarjala/19/orig 2025-09-07T07:36:04.1103809Z * [new branch] gh/slayton58/1/base -> origin/gh/slayton58/1/base 2025-09-07T07:36:04.1105021Z * [new branch] gh/slayton58/1/head -> origin/gh/slayton58/1/head 2025-09-07T07:36:04.1106214Z * [new branch] gh/slayton58/1/orig -> origin/gh/slayton58/1/orig 2025-09-07T07:36:04.1107896Z * [new branch] gh/slayton58/2/base -> origin/gh/slayton58/2/base 2025-09-07T07:36:04.1109187Z * [new branch] gh/slayton58/2/head -> origin/gh/slayton58/2/head 2025-09-07T07:36:04.1110466Z * [new branch] gh/slayton58/2/orig -> origin/gh/slayton58/2/orig 2025-09-07T07:36:04.1112155Z * [new branch] gh/slayton58/3/base -> origin/gh/slayton58/3/base 2025-09-07T07:36:04.1113280Z * [new branch] gh/slayton58/3/head -> origin/gh/slayton58/3/head 2025-09-07T07:36:04.1114476Z * [new branch] gh/slayton58/3/orig -> origin/gh/slayton58/3/orig 2025-09-07T07:36:04.1116168Z * [new branch] gh/slayton58/4/base -> origin/gh/slayton58/4/base 2025-09-07T07:36:04.1117306Z * [new branch] gh/slayton58/4/head -> origin/gh/slayton58/4/head 2025-09-07T07:36:04.1118512Z * [new branch] gh/slayton58/4/orig -> origin/gh/slayton58/4/orig 2025-09-07T07:36:04.1120219Z * [new branch] gh/slayton58/5/base -> origin/gh/slayton58/5/base 2025-09-07T07:36:04.1121435Z * [new branch] gh/slayton58/5/head -> origin/gh/slayton58/5/head 2025-09-07T07:36:04.1122642Z * [new branch] gh/slayton58/5/orig -> origin/gh/slayton58/5/orig 2025-09-07T07:36:04.1125047Z * [new branch] gh/soulitzer/269/base -> origin/gh/soulitzer/269/base 2025-09-07T07:36:04.1126201Z * [new branch] gh/soulitzer/269/head -> origin/gh/soulitzer/269/head 2025-09-07T07:36:04.1127426Z * [new branch] gh/soulitzer/269/orig -> origin/gh/soulitzer/269/orig 2025-09-07T07:36:04.1129216Z * [new branch] gh/soulitzer/276/base -> origin/gh/soulitzer/276/base 2025-09-07T07:36:04.1130463Z * [new branch] gh/soulitzer/276/head -> origin/gh/soulitzer/276/head 2025-09-07T07:36:04.1131814Z * [new branch] gh/soulitzer/276/orig -> origin/gh/soulitzer/276/orig 2025-09-07T07:36:04.1133733Z * [new branch] gh/soulitzer/287/base -> origin/gh/soulitzer/287/base 2025-09-07T07:36:04.1134879Z * [new branch] gh/soulitzer/287/head -> origin/gh/soulitzer/287/head 2025-09-07T07:36:04.1136531Z * [new branch] gh/soulitzer/287/orig -> origin/gh/soulitzer/287/orig 2025-09-07T07:36:04.1138401Z * [new branch] gh/soulitzer/296/base -> origin/gh/soulitzer/296/base 2025-09-07T07:36:04.1139605Z * [new branch] gh/soulitzer/296/head -> origin/gh/soulitzer/296/head 2025-09-07T07:36:04.1140846Z * [new branch] gh/soulitzer/296/orig -> origin/gh/soulitzer/296/orig 2025-09-07T07:36:04.1142666Z * [new branch] gh/soulitzer/299/base -> origin/gh/soulitzer/299/base 2025-09-07T07:36:04.1143933Z * [new branch] gh/soulitzer/299/head -> origin/gh/soulitzer/299/head 2025-09-07T07:36:04.1145110Z * [new branch] gh/soulitzer/299/orig -> origin/gh/soulitzer/299/orig 2025-09-07T07:36:04.1146905Z * [new branch] gh/soulitzer/300/base -> origin/gh/soulitzer/300/base 2025-09-07T07:36:04.1148065Z * [new branch] gh/soulitzer/300/head -> origin/gh/soulitzer/300/head 2025-09-07T07:36:04.1149248Z * [new branch] gh/soulitzer/300/orig -> origin/gh/soulitzer/300/orig 2025-09-07T07:36:04.1151053Z * [new branch] gh/soulitzer/301/base -> origin/gh/soulitzer/301/base 2025-09-07T07:36:04.1152292Z * [new branch] gh/soulitzer/301/head -> origin/gh/soulitzer/301/head 2025-09-07T07:36:04.1153479Z * [new branch] gh/soulitzer/301/orig -> origin/gh/soulitzer/301/orig 2025-09-07T07:36:04.1155177Z * [new branch] gh/soulitzer/313/base -> origin/gh/soulitzer/313/base 2025-09-07T07:36:04.1156359Z * [new branch] gh/soulitzer/313/head -> origin/gh/soulitzer/313/head 2025-09-07T07:36:04.1157538Z * [new branch] gh/soulitzer/313/orig -> origin/gh/soulitzer/313/orig 2025-09-07T07:36:04.1159861Z * [new branch] gh/soulitzer/319/base -> origin/gh/soulitzer/319/base 2025-09-07T07:36:04.1161079Z * [new branch] gh/soulitzer/319/head -> origin/gh/soulitzer/319/head 2025-09-07T07:36:04.1162272Z * [new branch] gh/soulitzer/319/orig -> origin/gh/soulitzer/319/orig 2025-09-07T07:36:04.1164097Z * [new branch] gh/soulitzer/320/base -> origin/gh/soulitzer/320/base 2025-09-07T07:36:04.1165222Z * [new branch] gh/soulitzer/320/head -> origin/gh/soulitzer/320/head 2025-09-07T07:36:04.1166411Z * [new branch] gh/soulitzer/320/orig -> origin/gh/soulitzer/320/orig 2025-09-07T07:36:04.1168171Z * [new branch] gh/soulitzer/336/base -> origin/gh/soulitzer/336/base 2025-09-07T07:36:04.1169462Z * [new branch] gh/soulitzer/336/head -> origin/gh/soulitzer/336/head 2025-09-07T07:36:04.1170643Z * [new branch] gh/soulitzer/336/orig -> origin/gh/soulitzer/336/orig 2025-09-07T07:36:04.1172358Z * [new branch] gh/soulitzer/347/base -> origin/gh/soulitzer/347/base 2025-09-07T07:36:04.1173491Z * [new branch] gh/soulitzer/347/head -> origin/gh/soulitzer/347/head 2025-09-07T07:36:04.1174661Z * [new branch] gh/soulitzer/347/orig -> origin/gh/soulitzer/347/orig 2025-09-07T07:36:04.1176605Z * [new branch] gh/soulitzer/349/base -> origin/gh/soulitzer/349/base 2025-09-07T07:36:04.1177924Z * [new branch] gh/soulitzer/349/head -> origin/gh/soulitzer/349/head 2025-09-07T07:36:04.1179200Z * [new branch] gh/soulitzer/349/orig -> origin/gh/soulitzer/349/orig 2025-09-07T07:36:04.1180843Z * [new branch] gh/soulitzer/350/base -> origin/gh/soulitzer/350/base 2025-09-07T07:36:04.1181956Z * [new branch] gh/soulitzer/350/head -> origin/gh/soulitzer/350/head 2025-09-07T07:36:04.1183154Z * [new branch] gh/soulitzer/350/orig -> origin/gh/soulitzer/350/orig 2025-09-07T07:36:04.1185344Z * [new branch] gh/soulitzer/351/base -> origin/gh/soulitzer/351/base 2025-09-07T07:36:04.1186530Z * [new branch] gh/soulitzer/351/head -> origin/gh/soulitzer/351/head 2025-09-07T07:36:04.1187709Z * [new branch] gh/soulitzer/351/orig -> origin/gh/soulitzer/351/orig 2025-09-07T07:36:04.1189386Z * [new branch] gh/soulitzer/353/base -> origin/gh/soulitzer/353/base 2025-09-07T07:36:04.1190680Z * [new branch] gh/soulitzer/353/head -> origin/gh/soulitzer/353/head 2025-09-07T07:36:04.1191860Z * [new branch] gh/soulitzer/353/orig -> origin/gh/soulitzer/353/orig 2025-09-07T07:36:04.1194183Z * [new branch] gh/soulitzer/358/base -> origin/gh/soulitzer/358/base 2025-09-07T07:36:04.1195462Z * [new branch] gh/soulitzer/358/head -> origin/gh/soulitzer/358/head 2025-09-07T07:36:04.1196732Z * [new branch] gh/soulitzer/358/orig -> origin/gh/soulitzer/358/orig 2025-09-07T07:36:04.1198800Z * [new branch] gh/soulitzer/359/base -> origin/gh/soulitzer/359/base 2025-09-07T07:36:04.1200008Z * [new branch] gh/soulitzer/359/head -> origin/gh/soulitzer/359/head 2025-09-07T07:36:04.1201215Z * [new branch] gh/soulitzer/359/orig -> origin/gh/soulitzer/359/orig 2025-09-07T07:36:04.1202957Z * [new branch] gh/soulitzer/362/base -> origin/gh/soulitzer/362/base 2025-09-07T07:36:04.1204394Z * [new branch] gh/soulitzer/362/head -> origin/gh/soulitzer/362/head 2025-09-07T07:36:04.1205828Z * [new branch] gh/soulitzer/362/orig -> origin/gh/soulitzer/362/orig 2025-09-07T07:36:04.1207591Z * [new branch] gh/soulitzer/372/base -> origin/gh/soulitzer/372/base 2025-09-07T07:36:04.1208759Z * [new branch] gh/soulitzer/372/head -> origin/gh/soulitzer/372/head 2025-09-07T07:36:04.1209926Z * [new branch] gh/soulitzer/372/orig -> origin/gh/soulitzer/372/orig 2025-09-07T07:36:04.1211877Z * [new branch] gh/soulitzer/373/base -> origin/gh/soulitzer/373/base 2025-09-07T07:36:04.1213038Z * [new branch] gh/soulitzer/373/head -> origin/gh/soulitzer/373/head 2025-09-07T07:36:04.1214238Z * [new branch] gh/soulitzer/373/orig -> origin/gh/soulitzer/373/orig 2025-09-07T07:36:04.1216047Z * [new branch] gh/soulitzer/374/base -> origin/gh/soulitzer/374/base 2025-09-07T07:36:04.1217320Z * [new branch] gh/soulitzer/374/head -> origin/gh/soulitzer/374/head 2025-09-07T07:36:04.1218561Z * [new branch] gh/soulitzer/374/orig -> origin/gh/soulitzer/374/orig 2025-09-07T07:36:04.1220323Z * [new branch] gh/soulitzer/375/base -> origin/gh/soulitzer/375/base 2025-09-07T07:36:04.1221456Z * [new branch] gh/soulitzer/375/head -> origin/gh/soulitzer/375/head 2025-09-07T07:36:04.1222651Z * [new branch] gh/soulitzer/375/orig -> origin/gh/soulitzer/375/orig 2025-09-07T07:36:04.1224374Z * [new branch] gh/soulitzer/376/base -> origin/gh/soulitzer/376/base 2025-09-07T07:36:04.1225590Z * [new branch] gh/soulitzer/376/head -> origin/gh/soulitzer/376/head 2025-09-07T07:36:04.1226769Z * [new branch] gh/soulitzer/376/orig -> origin/gh/soulitzer/376/orig 2025-09-07T07:36:04.1228682Z * [new branch] gh/soulitzer/377/base -> origin/gh/soulitzer/377/base 2025-09-07T07:36:04.1230249Z * [new branch] gh/soulitzer/377/head -> origin/gh/soulitzer/377/head 2025-09-07T07:36:04.1231559Z * [new branch] gh/soulitzer/377/orig -> origin/gh/soulitzer/377/orig 2025-09-07T07:36:04.1233371Z * [new branch] gh/soulitzer/378/base -> origin/gh/soulitzer/378/base 2025-09-07T07:36:04.1234564Z * [new branch] gh/soulitzer/378/head -> origin/gh/soulitzer/378/head 2025-09-07T07:36:04.1235824Z * [new branch] gh/soulitzer/378/orig -> origin/gh/soulitzer/378/orig 2025-09-07T07:36:04.1237584Z * [new branch] gh/soulitzer/379/base -> origin/gh/soulitzer/379/base 2025-09-07T07:36:04.1238847Z * [new branch] gh/soulitzer/379/head -> origin/gh/soulitzer/379/head 2025-09-07T07:36:04.1240029Z * [new branch] gh/soulitzer/379/orig -> origin/gh/soulitzer/379/orig 2025-09-07T07:36:04.1242857Z * [new branch] gh/swolchok/728/next -> origin/gh/swolchok/728/next 2025-09-07T07:36:04.1244762Z * [new branch] gh/swolchok/767/base -> origin/gh/swolchok/767/base 2025-09-07T07:36:04.1246276Z * [new branch] gh/swolchok/767/head -> origin/gh/swolchok/767/head 2025-09-07T07:36:04.1247736Z * [new branch] gh/swolchok/767/orig -> origin/gh/swolchok/767/orig 2025-09-07T07:36:04.1249406Z * [new branch] gh/swolchok/768/base -> origin/gh/swolchok/768/base 2025-09-07T07:36:04.1250709Z * [new branch] gh/swolchok/768/head -> origin/gh/swolchok/768/head 2025-09-07T07:36:04.1251921Z * [new branch] gh/swolchok/768/orig -> origin/gh/swolchok/768/orig 2025-09-07T07:36:04.1253780Z * [new branch] gh/swolchok/769/base -> origin/gh/swolchok/769/base 2025-09-07T07:36:04.1255118Z * [new branch] gh/swolchok/769/head -> origin/gh/swolchok/769/head 2025-09-07T07:36:04.1256520Z * [new branch] gh/swolchok/769/orig -> origin/gh/swolchok/769/orig 2025-09-07T07:36:04.1258280Z * [new branch] gh/swolchok/771/base -> origin/gh/swolchok/771/base 2025-09-07T07:36:04.1259659Z * [new branch] gh/swolchok/771/head -> origin/gh/swolchok/771/head 2025-09-07T07:36:04.1260951Z * [new branch] gh/swolchok/771/orig -> origin/gh/swolchok/771/orig 2025-09-07T07:36:04.1262878Z * [new branch] gh/swolchok/772/base -> origin/gh/swolchok/772/base 2025-09-07T07:36:04.1264164Z * [new branch] gh/swolchok/772/head -> origin/gh/swolchok/772/head 2025-09-07T07:36:04.1265405Z * [new branch] gh/swolchok/772/orig -> origin/gh/swolchok/772/orig 2025-09-07T07:36:04.1267257Z * [new branch] gh/swolchok/773/base -> origin/gh/swolchok/773/base 2025-09-07T07:36:04.1268458Z * [new branch] gh/swolchok/773/head -> origin/gh/swolchok/773/head 2025-09-07T07:36:04.1269687Z * [new branch] gh/swolchok/773/orig -> origin/gh/swolchok/773/orig 2025-09-07T07:36:04.1271440Z * [new branch] gh/swolchok/786/base -> origin/gh/swolchok/786/base 2025-09-07T07:36:04.1272602Z * [new branch] gh/swolchok/786/head -> origin/gh/swolchok/786/head 2025-09-07T07:36:04.1273860Z * [new branch] gh/swolchok/786/orig -> origin/gh/swolchok/786/orig 2025-09-07T07:36:04.1275461Z * [new branch] gh/swolchok/787/base -> origin/gh/swolchok/787/base 2025-09-07T07:36:04.1276632Z * [new branch] gh/swolchok/787/head -> origin/gh/swolchok/787/head 2025-09-07T07:36:04.1277877Z * [new branch] gh/swolchok/787/orig -> origin/gh/swolchok/787/orig 2025-09-07T07:36:04.1279744Z * [new branch] gh/swolchok/788/base -> origin/gh/swolchok/788/base 2025-09-07T07:36:04.1281019Z * [new branch] gh/swolchok/788/head -> origin/gh/swolchok/788/head 2025-09-07T07:36:04.1282186Z * [new branch] gh/swolchok/788/orig -> origin/gh/swolchok/788/orig 2025-09-07T07:36:04.1283861Z * [new branch] gh/swolchok/789/base -> origin/gh/swolchok/789/base 2025-09-07T07:36:04.1285021Z * [new branch] gh/swolchok/789/head -> origin/gh/swolchok/789/head 2025-09-07T07:36:04.1286305Z * [new branch] gh/swolchok/789/orig -> origin/gh/swolchok/789/orig 2025-09-07T07:36:04.1287989Z * [new branch] gh/swolchok/790/base -> origin/gh/swolchok/790/base 2025-09-07T07:36:04.1289199Z * [new branch] gh/swolchok/790/head -> origin/gh/swolchok/790/head 2025-09-07T07:36:04.1290384Z * [new branch] gh/swolchok/790/orig -> origin/gh/swolchok/790/orig 2025-09-07T07:36:04.1292172Z * [new branch] gh/swolchok/791/base -> origin/gh/swolchok/791/base 2025-09-07T07:36:04.1293278Z * [new branch] gh/swolchok/791/head -> origin/gh/swolchok/791/head 2025-09-07T07:36:04.1294532Z * [new branch] gh/swolchok/791/orig -> origin/gh/swolchok/791/orig 2025-09-07T07:36:04.1296240Z * [new branch] gh/swolchok/792/base -> origin/gh/swolchok/792/base 2025-09-07T07:36:04.1297753Z * [new branch] gh/swolchok/792/head -> origin/gh/swolchok/792/head 2025-09-07T07:36:04.1299204Z * [new branch] gh/swolchok/792/orig -> origin/gh/swolchok/792/orig 2025-09-07T07:36:04.1300986Z * [new branch] gh/swolchok/793/base -> origin/gh/swolchok/793/base 2025-09-07T07:36:04.1302124Z * [new branch] gh/swolchok/793/head -> origin/gh/swolchok/793/head 2025-09-07T07:36:04.1303558Z * [new branch] gh/swolchok/793/orig -> origin/gh/swolchok/793/orig 2025-09-07T07:36:04.1305363Z * [new branch] gh/swolchok/794/base -> origin/gh/swolchok/794/base 2025-09-07T07:36:04.1306543Z * [new branch] gh/swolchok/794/head -> origin/gh/swolchok/794/head 2025-09-07T07:36:04.1307664Z * [new branch] gh/swolchok/794/orig -> origin/gh/swolchok/794/orig 2025-09-07T07:36:04.1309856Z * [new branch] gh/swolchok/795/base -> origin/gh/swolchok/795/base 2025-09-07T07:36:04.1311091Z * [new branch] gh/swolchok/795/head -> origin/gh/swolchok/795/head 2025-09-07T07:36:04.1312330Z * [new branch] gh/swolchok/795/orig -> origin/gh/swolchok/795/orig 2025-09-07T07:36:04.1314098Z * [new branch] gh/swolchok/796/base -> origin/gh/swolchok/796/base 2025-09-07T07:36:04.1315620Z * [new branch] gh/swolchok/796/head -> origin/gh/swolchok/796/head 2025-09-07T07:36:04.1316815Z * [new branch] gh/swolchok/796/orig -> origin/gh/swolchok/796/orig 2025-09-07T07:36:04.1318849Z * [new branch] gh/swolchok/797/base -> origin/gh/swolchok/797/base 2025-09-07T07:36:04.1320591Z * [new branch] gh/swolchok/797/head -> origin/gh/swolchok/797/head 2025-09-07T07:36:04.1321871Z * [new branch] gh/swolchok/797/orig -> origin/gh/swolchok/797/orig 2025-09-07T07:36:04.1323688Z * [new branch] gh/swolchok/798/base -> origin/gh/swolchok/798/base 2025-09-07T07:36:04.1324812Z * [new branch] gh/swolchok/798/head -> origin/gh/swolchok/798/head 2025-09-07T07:36:04.1326098Z * [new branch] gh/swolchok/798/orig -> origin/gh/swolchok/798/orig 2025-09-07T07:36:04.1328404Z * [new branch] gh/swolchok/799/base -> origin/gh/swolchok/799/base 2025-09-07T07:36:04.1329544Z * [new branch] gh/swolchok/799/head -> origin/gh/swolchok/799/head 2025-09-07T07:36:04.1330863Z * [new branch] gh/swolchok/799/orig -> origin/gh/swolchok/799/orig 2025-09-07T07:36:04.1332760Z * [new branch] gh/swolchok/800/base -> origin/gh/swolchok/800/base 2025-09-07T07:36:04.1333974Z * [new branch] gh/swolchok/800/head -> origin/gh/swolchok/800/head 2025-09-07T07:36:04.1335260Z * [new branch] gh/swolchok/800/orig -> origin/gh/swolchok/800/orig 2025-09-07T07:36:04.1337093Z * [new branch] gh/swolchok/801/base -> origin/gh/swolchok/801/base 2025-09-07T07:36:04.1338341Z * [new branch] gh/swolchok/801/head -> origin/gh/swolchok/801/head 2025-09-07T07:36:04.1339658Z * [new branch] gh/swolchok/801/orig -> origin/gh/swolchok/801/orig 2025-09-07T07:36:04.1341429Z * [new branch] gh/swolchok/802/base -> origin/gh/swolchok/802/base 2025-09-07T07:36:04.1342548Z * [new branch] gh/swolchok/802/head -> origin/gh/swolchok/802/head 2025-09-07T07:36:04.1343792Z * [new branch] gh/swolchok/802/orig -> origin/gh/swolchok/802/orig 2025-09-07T07:36:04.1345541Z * [new branch] gh/swolchok/803/base -> origin/gh/swolchok/803/base 2025-09-07T07:36:04.1346733Z * [new branch] gh/swolchok/803/head -> origin/gh/swolchok/803/head 2025-09-07T07:36:04.1348001Z * [new branch] gh/swolchok/803/orig -> origin/gh/swolchok/803/orig 2025-09-07T07:36:04.1349927Z * [new branch] gh/swolchok/804/base -> origin/gh/swolchok/804/base 2025-09-07T07:36:04.1351001Z * [new branch] gh/swolchok/804/head -> origin/gh/swolchok/804/head 2025-09-07T07:36:04.1352403Z * [new branch] gh/swolchok/804/orig -> origin/gh/swolchok/804/orig 2025-09-07T07:36:04.1354190Z * [new branch] gh/swolchok/805/base -> origin/gh/swolchok/805/base 2025-09-07T07:36:04.1355398Z * [new branch] gh/swolchok/805/head -> origin/gh/swolchok/805/head 2025-09-07T07:36:04.1356717Z * [new branch] gh/swolchok/805/orig -> origin/gh/swolchok/805/orig 2025-09-07T07:36:04.1358386Z * [new branch] gh/swolchok/806/base -> origin/gh/swolchok/806/base 2025-09-07T07:36:04.1359626Z * [new branch] gh/swolchok/806/head -> origin/gh/swolchok/806/head 2025-09-07T07:36:04.1360929Z * [new branch] gh/swolchok/806/orig -> origin/gh/swolchok/806/orig 2025-09-07T07:36:04.1362730Z * [new branch] gh/swolchok/807/base -> origin/gh/swolchok/807/base 2025-09-07T07:36:04.1363984Z * [new branch] gh/swolchok/807/head -> origin/gh/swolchok/807/head 2025-09-07T07:36:04.1365421Z * [new branch] gh/swolchok/807/orig -> origin/gh/swolchok/807/orig 2025-09-07T07:36:04.1367459Z * [new branch] gh/swolchok/808/base -> origin/gh/swolchok/808/base 2025-09-07T07:36:04.1368769Z * [new branch] gh/swolchok/808/head -> origin/gh/swolchok/808/head 2025-09-07T07:36:04.1369960Z * [new branch] gh/swolchok/808/orig -> origin/gh/swolchok/808/orig 2025-09-07T07:36:04.1371759Z * [new branch] gh/swolchok/809/base -> origin/gh/swolchok/809/base 2025-09-07T07:36:04.1372947Z * [new branch] gh/swolchok/809/head -> origin/gh/swolchok/809/head 2025-09-07T07:36:04.1374149Z * [new branch] gh/swolchok/809/orig -> origin/gh/swolchok/809/orig 2025-09-07T07:36:04.1376063Z * [new branch] gh/swolchok/810/base -> origin/gh/swolchok/810/base 2025-09-07T07:36:04.1377186Z * [new branch] gh/swolchok/810/head -> origin/gh/swolchok/810/head 2025-09-07T07:36:04.1378491Z * [new branch] gh/swolchok/810/orig -> origin/gh/swolchok/810/orig 2025-09-07T07:36:04.1380285Z * [new branch] gh/swolchok/811/base -> origin/gh/swolchok/811/base 2025-09-07T07:36:04.1381555Z * [new branch] gh/swolchok/811/head -> origin/gh/swolchok/811/head 2025-09-07T07:36:04.1382784Z * [new branch] gh/swolchok/811/orig -> origin/gh/swolchok/811/orig 2025-09-07T07:36:04.1384598Z * [new branch] gh/swolchok/812/base -> origin/gh/swolchok/812/base 2025-09-07T07:36:04.1385778Z * [new branch] gh/swolchok/812/head -> origin/gh/swolchok/812/head 2025-09-07T07:36:04.1387414Z * [new branch] gh/swolchok/812/orig -> origin/gh/swolchok/812/orig 2025-09-07T07:36:04.1389362Z * [new branch] gh/swolchok/813/base -> origin/gh/swolchok/813/base 2025-09-07T07:36:04.1390684Z * [new branch] gh/swolchok/813/head -> origin/gh/swolchok/813/head 2025-09-07T07:36:04.1391876Z * [new branch] gh/swolchok/813/orig -> origin/gh/swolchok/813/orig 2025-09-07T07:36:04.1393756Z * [new branch] gh/swolchok/814/base -> origin/gh/swolchok/814/base 2025-09-07T07:36:04.1394881Z * [new branch] gh/swolchok/814/head -> origin/gh/swolchok/814/head 2025-09-07T07:36:04.1396073Z * [new branch] gh/swolchok/814/orig -> origin/gh/swolchok/814/orig 2025-09-07T07:36:04.1397973Z * [new branch] gh/swolchok/815/base -> origin/gh/swolchok/815/base 2025-09-07T07:36:04.1399141Z * [new branch] gh/swolchok/815/head -> origin/gh/swolchok/815/head 2025-09-07T07:36:04.1400483Z * [new branch] gh/swolchok/815/orig -> origin/gh/swolchok/815/orig 2025-09-07T07:36:04.1402224Z * [new branch] gh/swolchok/816/base -> origin/gh/swolchok/816/base 2025-09-07T07:36:04.1403712Z * [new branch] gh/swolchok/816/head -> origin/gh/swolchok/816/head 2025-09-07T07:36:04.1405414Z * [new branch] gh/swolchok/816/orig -> origin/gh/swolchok/816/orig 2025-09-07T07:36:04.1407298Z * [new branch] gh/swolchok/817/base -> origin/gh/swolchok/817/base 2025-09-07T07:36:04.1408514Z * [new branch] gh/swolchok/817/head -> origin/gh/swolchok/817/head 2025-09-07T07:36:04.1409636Z * [new branch] gh/swolchok/817/orig -> origin/gh/swolchok/817/orig 2025-09-07T07:36:04.1411423Z * [new branch] gh/swolchok/818/base -> origin/gh/swolchok/818/base 2025-09-07T07:36:04.1412539Z * [new branch] gh/swolchok/818/head -> origin/gh/swolchok/818/head 2025-09-07T07:36:04.1413860Z * [new branch] gh/swolchok/818/orig -> origin/gh/swolchok/818/orig 2025-09-07T07:36:04.1416102Z * [new branch] gh/swolchok/819/base -> origin/gh/swolchok/819/base 2025-09-07T07:36:04.1417245Z * [new branch] gh/swolchok/819/head -> origin/gh/swolchok/819/head 2025-09-07T07:36:04.1418515Z * [new branch] gh/swolchok/819/orig -> origin/gh/swolchok/819/orig 2025-09-07T07:36:04.1420328Z * [new branch] gh/swolchok/820/base -> origin/gh/swolchok/820/base 2025-09-07T07:36:04.1421565Z * [new branch] gh/swolchok/820/head -> origin/gh/swolchok/820/head 2025-09-07T07:36:04.1423146Z * [new branch] gh/swolchok/820/orig -> origin/gh/swolchok/820/orig 2025-09-07T07:36:04.1424990Z * [new branch] gh/swolchok/821/base -> origin/gh/swolchok/821/base 2025-09-07T07:36:04.1426137Z * [new branch] gh/swolchok/821/head -> origin/gh/swolchok/821/head 2025-09-07T07:36:04.1427511Z * [new branch] gh/swolchok/821/orig -> origin/gh/swolchok/821/orig 2025-09-07T07:36:04.1429409Z * [new branch] gh/swolchok/822/base -> origin/gh/swolchok/822/base 2025-09-07T07:36:04.1430562Z * [new branch] gh/swolchok/822/head -> origin/gh/swolchok/822/head 2025-09-07T07:36:04.1431746Z * [new branch] gh/swolchok/822/orig -> origin/gh/swolchok/822/orig 2025-09-07T07:36:04.1433713Z * [new branch] gh/swolchok/823/base -> origin/gh/swolchok/823/base 2025-09-07T07:36:04.1434858Z * [new branch] gh/swolchok/823/head -> origin/gh/swolchok/823/head 2025-09-07T07:36:04.1436063Z * [new branch] gh/swolchok/823/orig -> origin/gh/swolchok/823/orig 2025-09-07T07:36:04.1437797Z * [new branch] gh/swolchok/824/base -> origin/gh/swolchok/824/base 2025-09-07T07:36:04.1438990Z * [new branch] gh/swolchok/824/head -> origin/gh/swolchok/824/head 2025-09-07T07:36:04.1440163Z * [new branch] gh/swolchok/824/orig -> origin/gh/swolchok/824/orig 2025-09-07T07:36:04.1441938Z * [new branch] gh/swolchok/825/base -> origin/gh/swolchok/825/base 2025-09-07T07:36:04.1443181Z * [new branch] gh/swolchok/825/head -> origin/gh/swolchok/825/head 2025-09-07T07:36:04.1444369Z * [new branch] gh/swolchok/825/orig -> origin/gh/swolchok/825/orig 2025-09-07T07:36:04.1446199Z * [new branch] gh/swolchok/826/base -> origin/gh/swolchok/826/base 2025-09-07T07:36:04.1447364Z * [new branch] gh/swolchok/826/head -> origin/gh/swolchok/826/head 2025-09-07T07:36:04.1448448Z * [new branch] gh/swolchok/826/orig -> origin/gh/swolchok/826/orig 2025-09-07T07:36:04.1450311Z * [new branch] gh/swolchok/827/base -> origin/gh/swolchok/827/base 2025-09-07T07:36:04.1451571Z * [new branch] gh/swolchok/827/head -> origin/gh/swolchok/827/head 2025-09-07T07:36:04.1452601Z * [new branch] gh/swolchok/827/orig -> origin/gh/swolchok/827/orig 2025-09-07T07:36:04.1454427Z * [new branch] gh/swolchok/828/base -> origin/gh/swolchok/828/base 2025-09-07T07:36:04.1455614Z * [new branch] gh/swolchok/828/head -> origin/gh/swolchok/828/head 2025-09-07T07:36:04.1456792Z * [new branch] gh/swolchok/828/orig -> origin/gh/swolchok/828/orig 2025-09-07T07:36:04.1458518Z * [new branch] gh/swolchok/829/base -> origin/gh/swolchok/829/base 2025-09-07T07:36:04.1459726Z * [new branch] gh/swolchok/829/head -> origin/gh/swolchok/829/head 2025-09-07T07:36:04.1460916Z * [new branch] gh/swolchok/829/orig -> origin/gh/swolchok/829/orig 2025-09-07T07:36:04.1462906Z * [new branch] gh/swolchok/830/base -> origin/gh/swolchok/830/base 2025-09-07T07:36:04.1464067Z * [new branch] gh/swolchok/830/head -> origin/gh/swolchok/830/head 2025-09-07T07:36:04.1465214Z * [new branch] gh/swolchok/830/orig -> origin/gh/swolchok/830/orig 2025-09-07T07:36:04.1466835Z * [new branch] gh/swolchok/831/base -> origin/gh/swolchok/831/base 2025-09-07T07:36:04.1468140Z * [new branch] gh/swolchok/831/head -> origin/gh/swolchok/831/head 2025-09-07T07:36:04.1469440Z * [new branch] gh/swolchok/831/orig -> origin/gh/swolchok/831/orig 2025-09-07T07:36:04.1471868Z * [new branch] gh/swolchok/832/base -> origin/gh/swolchok/832/base 2025-09-07T07:36:04.1473203Z * [new branch] gh/swolchok/832/head -> origin/gh/swolchok/832/head 2025-09-07T07:36:04.1474395Z * [new branch] gh/swolchok/832/orig -> origin/gh/swolchok/832/orig 2025-09-07T07:36:04.1476560Z * [new branch] gh/syed-ahmed/3/base -> origin/gh/syed-ahmed/3/base 2025-09-07T07:36:04.1477733Z * [new branch] gh/syed-ahmed/3/head -> origin/gh/syed-ahmed/3/head 2025-09-07T07:36:04.1478959Z * [new branch] gh/syed-ahmed/3/orig -> origin/gh/syed-ahmed/3/orig 2025-09-07T07:36:04.1480723Z * [new branch] gh/syed-ahmed/4/base -> origin/gh/syed-ahmed/4/base 2025-09-07T07:36:04.1481874Z * [new branch] gh/syed-ahmed/4/head -> origin/gh/syed-ahmed/4/head 2025-09-07T07:36:04.1483061Z * [new branch] gh/syed-ahmed/4/orig -> origin/gh/syed-ahmed/4/orig 2025-09-07T07:36:04.1484732Z * [new branch] gh/syed-ahmed/5/base -> origin/gh/syed-ahmed/5/base 2025-09-07T07:36:04.1485915Z * [new branch] gh/syed-ahmed/5/head -> origin/gh/syed-ahmed/5/head 2025-09-07T07:36:04.1487078Z * [new branch] gh/syed-ahmed/5/orig -> origin/gh/syed-ahmed/5/orig 2025-09-07T07:36:04.1489379Z * [new branch] gh/teja-rao/4/base -> origin/gh/teja-rao/4/base 2025-09-07T07:36:04.1490609Z * [new branch] gh/teja-rao/4/head -> origin/gh/teja-rao/4/head 2025-09-07T07:36:04.1491817Z * [new branch] gh/teja-rao/4/orig -> origin/gh/teja-rao/4/orig 2025-09-07T07:36:04.1493909Z * [new branch] gh/tianyu-l/2/base -> origin/gh/tianyu-l/2/base 2025-09-07T07:36:04.1495116Z * [new branch] gh/tianyu-l/2/head -> origin/gh/tianyu-l/2/head 2025-09-07T07:36:04.1496308Z * [new branch] gh/tianyu-l/2/orig -> origin/gh/tianyu-l/2/orig 2025-09-07T07:36:04.1498092Z * [new branch] gh/tianyu-l/3/base -> origin/gh/tianyu-l/3/base 2025-09-07T07:36:04.1499439Z * [new branch] gh/tianyu-l/3/head -> origin/gh/tianyu-l/3/head 2025-09-07T07:36:04.1500656Z * [new branch] gh/tianyu-l/3/orig -> origin/gh/tianyu-l/3/orig 2025-09-07T07:36:04.1502405Z * [new branch] gh/tianyu-l/4/base -> origin/gh/tianyu-l/4/base 2025-09-07T07:36:04.1503931Z * [new branch] gh/tianyu-l/4/head -> origin/gh/tianyu-l/4/head 2025-09-07T07:36:04.1505055Z * [new branch] gh/tianyu-l/4/orig -> origin/gh/tianyu-l/4/orig 2025-09-07T07:36:04.1507284Z * [new branch] gh/tugsbayasgalan/1/base -> origin/gh/tugsbayasgalan/1/base 2025-09-07T07:36:04.1508425Z * [new branch] gh/tugsbayasgalan/1/head -> origin/gh/tugsbayasgalan/1/head 2025-09-07T07:36:04.1509751Z * [new branch] gh/tugsbayasgalan/1/orig -> origin/gh/tugsbayasgalan/1/orig 2025-09-07T07:36:04.1511699Z * [new branch] gh/tugsbayasgalan/10/base -> origin/gh/tugsbayasgalan/10/base 2025-09-07T07:36:04.1512896Z * [new branch] gh/tugsbayasgalan/10/head -> origin/gh/tugsbayasgalan/10/head 2025-09-07T07:36:04.1514074Z * [new branch] gh/tugsbayasgalan/10/orig -> origin/gh/tugsbayasgalan/10/orig 2025-09-07T07:36:04.1515749Z * [new branch] gh/tugsbayasgalan/11/base -> origin/gh/tugsbayasgalan/11/base 2025-09-07T07:36:04.1516979Z * [new branch] gh/tugsbayasgalan/11/head -> origin/gh/tugsbayasgalan/11/head 2025-09-07T07:36:04.1518145Z * [new branch] gh/tugsbayasgalan/11/orig -> origin/gh/tugsbayasgalan/11/orig 2025-09-07T07:36:04.1519914Z * [new branch] gh/tugsbayasgalan/12/base -> origin/gh/tugsbayasgalan/12/base 2025-09-07T07:36:04.1521116Z * [new branch] gh/tugsbayasgalan/12/head -> origin/gh/tugsbayasgalan/12/head 2025-09-07T07:36:04.1522316Z * [new branch] gh/tugsbayasgalan/12/orig -> origin/gh/tugsbayasgalan/12/orig 2025-09-07T07:36:04.1524117Z * [new branch] gh/tugsbayasgalan/13/base -> origin/gh/tugsbayasgalan/13/base 2025-09-07T07:36:04.1525317Z * [new branch] gh/tugsbayasgalan/13/head -> origin/gh/tugsbayasgalan/13/head 2025-09-07T07:36:04.1526554Z * [new branch] gh/tugsbayasgalan/13/orig -> origin/gh/tugsbayasgalan/13/orig 2025-09-07T07:36:04.1528797Z * [new branch] gh/tugsbayasgalan/14/base -> origin/gh/tugsbayasgalan/14/base 2025-09-07T07:36:04.1529948Z * [new branch] gh/tugsbayasgalan/14/head -> origin/gh/tugsbayasgalan/14/head 2025-09-07T07:36:04.1531117Z * [new branch] gh/tugsbayasgalan/14/orig -> origin/gh/tugsbayasgalan/14/orig 2025-09-07T07:36:04.1532983Z * [new branch] gh/tugsbayasgalan/15/base -> origin/gh/tugsbayasgalan/15/base 2025-09-07T07:36:04.1534124Z * [new branch] gh/tugsbayasgalan/15/head -> origin/gh/tugsbayasgalan/15/head 2025-09-07T07:36:04.1535435Z * [new branch] gh/tugsbayasgalan/15/orig -> origin/gh/tugsbayasgalan/15/orig 2025-09-07T07:36:04.1537159Z * [new branch] gh/tugsbayasgalan/2/base -> origin/gh/tugsbayasgalan/2/base 2025-09-07T07:36:04.1538534Z * [new branch] gh/tugsbayasgalan/2/head -> origin/gh/tugsbayasgalan/2/head 2025-09-07T07:36:04.1539730Z * [new branch] gh/tugsbayasgalan/2/orig -> origin/gh/tugsbayasgalan/2/orig 2025-09-07T07:36:04.1541730Z * [new branch] gh/tugsbayasgalan/3/base -> origin/gh/tugsbayasgalan/3/base 2025-09-07T07:36:04.1543028Z * [new branch] gh/tugsbayasgalan/3/head -> origin/gh/tugsbayasgalan/3/head 2025-09-07T07:36:04.1544222Z * [new branch] gh/tugsbayasgalan/3/orig -> origin/gh/tugsbayasgalan/3/orig 2025-09-07T07:36:04.1545938Z * [new branch] gh/tugsbayasgalan/4/base -> origin/gh/tugsbayasgalan/4/base 2025-09-07T07:36:04.1547327Z * [new branch] gh/tugsbayasgalan/4/head -> origin/gh/tugsbayasgalan/4/head 2025-09-07T07:36:04.1548487Z * [new branch] gh/tugsbayasgalan/4/orig -> origin/gh/tugsbayasgalan/4/orig 2025-09-07T07:36:04.1550271Z * [new branch] gh/tugsbayasgalan/5/base -> origin/gh/tugsbayasgalan/5/base 2025-09-07T07:36:04.1551592Z * [new branch] gh/tugsbayasgalan/5/head -> origin/gh/tugsbayasgalan/5/head 2025-09-07T07:36:04.1552719Z * [new branch] gh/tugsbayasgalan/5/orig -> origin/gh/tugsbayasgalan/5/orig 2025-09-07T07:36:04.1554340Z * [new branch] gh/tugsbayasgalan/6/base -> origin/gh/tugsbayasgalan/6/base 2025-09-07T07:36:04.1555644Z * [new branch] gh/tugsbayasgalan/6/head -> origin/gh/tugsbayasgalan/6/head 2025-09-07T07:36:04.1556871Z * [new branch] gh/tugsbayasgalan/6/orig -> origin/gh/tugsbayasgalan/6/orig 2025-09-07T07:36:04.1558621Z * [new branch] gh/tugsbayasgalan/7/base -> origin/gh/tugsbayasgalan/7/base 2025-09-07T07:36:04.1559791Z * [new branch] gh/tugsbayasgalan/7/head -> origin/gh/tugsbayasgalan/7/head 2025-09-07T07:36:04.1561056Z * [new branch] gh/tugsbayasgalan/7/orig -> origin/gh/tugsbayasgalan/7/orig 2025-09-07T07:36:04.1562765Z * [new branch] gh/tugsbayasgalan/8/base -> origin/gh/tugsbayasgalan/8/base 2025-09-07T07:36:04.1563899Z * [new branch] gh/tugsbayasgalan/8/head -> origin/gh/tugsbayasgalan/8/head 2025-09-07T07:36:04.1565092Z * [new branch] gh/tugsbayasgalan/8/orig -> origin/gh/tugsbayasgalan/8/orig 2025-09-07T07:36:04.1566743Z * [new branch] gh/tugsbayasgalan/9/base -> origin/gh/tugsbayasgalan/9/base 2025-09-07T07:36:04.1567813Z * [new branch] gh/tugsbayasgalan/9/head -> origin/gh/tugsbayasgalan/9/head 2025-09-07T07:36:04.1568973Z * [new branch] gh/tugsbayasgalan/9/orig -> origin/gh/tugsbayasgalan/9/orig 2025-09-07T07:36:04.1571361Z * [new branch] gh/v0i0/1/base -> origin/gh/v0i0/1/base 2025-09-07T07:36:04.1572540Z * [new branch] gh/v0i0/1/head -> origin/gh/v0i0/1/head 2025-09-07T07:36:04.1573716Z * [new branch] gh/v0i0/1/orig -> origin/gh/v0i0/1/orig 2025-09-07T07:36:04.1575422Z * [new branch] gh/v0i0/4/base -> origin/gh/v0i0/4/base 2025-09-07T07:36:04.1576582Z * [new branch] gh/v0i0/4/head -> origin/gh/v0i0/4/head 2025-09-07T07:36:04.1577758Z * [new branch] gh/v0i0/4/orig -> origin/gh/v0i0/4/orig 2025-09-07T07:36:04.1579557Z * [new branch] gh/v0i0/6/base -> origin/gh/v0i0/6/base 2025-09-07T07:36:04.1580755Z * [new branch] gh/v0i0/6/head -> origin/gh/v0i0/6/head 2025-09-07T07:36:04.1581927Z * [new branch] gh/v0i0/6/orig -> origin/gh/v0i0/6/orig 2025-09-07T07:36:04.1583659Z * [new branch] gh/v0i0/7/base -> origin/gh/v0i0/7/base 2025-09-07T07:36:04.1584885Z * [new branch] gh/v0i0/7/head -> origin/gh/v0i0/7/head 2025-09-07T07:36:04.1586056Z * [new branch] gh/v0i0/7/orig -> origin/gh/v0i0/7/orig 2025-09-07T07:36:04.1587785Z * [new branch] gh/v0i0/8/base -> origin/gh/v0i0/8/base 2025-09-07T07:36:04.1588911Z * [new branch] gh/v0i0/8/head -> origin/gh/v0i0/8/head 2025-09-07T07:36:04.1590061Z * [new branch] gh/v0i0/8/orig -> origin/gh/v0i0/8/orig 2025-09-07T07:36:04.1591788Z * [new branch] gh/v0i0/9/base -> origin/gh/v0i0/9/base 2025-09-07T07:36:04.1592968Z * [new branch] gh/v0i0/9/head -> origin/gh/v0i0/9/head 2025-09-07T07:36:04.1594265Z * [new branch] gh/v0i0/9/orig -> origin/gh/v0i0/9/orig 2025-09-07T07:36:04.1596329Z * [new branch] gh/vkuzo/1/next -> origin/gh/vkuzo/1/next 2025-09-07T07:36:04.1597991Z * [new branch] gh/vkuzo/2/next -> origin/gh/vkuzo/2/next 2025-09-07T07:36:04.1599682Z * [new branch] gh/vkuzo/3/next -> origin/gh/vkuzo/3/next 2025-09-07T07:36:04.1601372Z * [new branch] gh/vkuzo/4/base -> origin/gh/vkuzo/4/base 2025-09-07T07:36:04.1602871Z * [new branch] gh/vkuzo/4/head -> origin/gh/vkuzo/4/head 2025-09-07T07:36:04.1606374Z * [new branch] gh/vkuzo/4/orig -> origin/gh/vkuzo/4/orig 2025-09-07T07:36:04.1608230Z * [new branch] gh/vkuzo/5/base -> origin/gh/vkuzo/5/base 2025-09-07T07:36:04.1609614Z * [new branch] gh/vkuzo/5/head -> origin/gh/vkuzo/5/head 2025-09-07T07:36:04.1611046Z * [new branch] gh/vkuzo/5/orig -> origin/gh/vkuzo/5/orig 2025-09-07T07:36:04.1612912Z * [new branch] gh/vkuzo/6/base -> origin/gh/vkuzo/6/base 2025-09-07T07:36:04.1614040Z * [new branch] gh/vkuzo/6/head -> origin/gh/vkuzo/6/head 2025-09-07T07:36:04.1615300Z * [new branch] gh/vkuzo/6/orig -> origin/gh/vkuzo/6/orig 2025-09-07T07:36:04.1616886Z * [new branch] gh/vkuzo/7/base -> origin/gh/vkuzo/7/base 2025-09-07T07:36:04.1618410Z * [new branch] gh/vkuzo/7/head -> origin/gh/vkuzo/7/head 2025-09-07T07:36:04.1619546Z * [new branch] gh/vkuzo/7/orig -> origin/gh/vkuzo/7/orig 2025-09-07T07:36:04.1621772Z * [new branch] gh/wconstab/419/base -> origin/gh/wconstab/419/base 2025-09-07T07:36:04.1622895Z * [new branch] gh/wconstab/419/head -> origin/gh/wconstab/419/head 2025-09-07T07:36:04.1624103Z * [new branch] gh/wconstab/419/orig -> origin/gh/wconstab/419/orig 2025-09-07T07:36:04.1625900Z * [new branch] gh/wconstab/424/base -> origin/gh/wconstab/424/base 2025-09-07T07:36:04.1627021Z * [new branch] gh/wconstab/424/head -> origin/gh/wconstab/424/head 2025-09-07T07:36:04.1628164Z * [new branch] gh/wconstab/424/orig -> origin/gh/wconstab/424/orig 2025-09-07T07:36:04.1629873Z * [new branch] gh/wconstab/435/base -> origin/gh/wconstab/435/base 2025-09-07T07:36:04.1631080Z * [new branch] gh/wconstab/435/head -> origin/gh/wconstab/435/head 2025-09-07T07:36:04.1632272Z * [new branch] gh/wconstab/435/orig -> origin/gh/wconstab/435/orig 2025-09-07T07:36:04.1633976Z * [new branch] gh/wconstab/438/base -> origin/gh/wconstab/438/base 2025-09-07T07:36:04.1635201Z * [new branch] gh/wconstab/438/head -> origin/gh/wconstab/438/head 2025-09-07T07:36:04.1636794Z * [new branch] gh/wconstab/438/orig -> origin/gh/wconstab/438/orig 2025-09-07T07:36:04.1639091Z * [new branch] gh/wconstab/440/base -> origin/gh/wconstab/440/base 2025-09-07T07:36:04.1640411Z * [new branch] gh/wconstab/440/head -> origin/gh/wconstab/440/head 2025-09-07T07:36:04.1641677Z * [new branch] gh/wconstab/440/orig -> origin/gh/wconstab/440/orig 2025-09-07T07:36:04.1643580Z * [new branch] gh/wconstab/441/base -> origin/gh/wconstab/441/base 2025-09-07T07:36:04.1644721Z * [new branch] gh/wconstab/441/head -> origin/gh/wconstab/441/head 2025-09-07T07:36:04.1645953Z * [new branch] gh/wconstab/441/orig -> origin/gh/wconstab/441/orig 2025-09-07T07:36:04.1647925Z * [new branch] gh/wconstab/442/base -> origin/gh/wconstab/442/base 2025-09-07T07:36:04.1649228Z * [new branch] gh/wconstab/442/head -> origin/gh/wconstab/442/head 2025-09-07T07:36:04.1650438Z * [new branch] gh/wconstab/442/orig -> origin/gh/wconstab/442/orig 2025-09-07T07:36:04.1652178Z * [new branch] gh/wconstab/443/base -> origin/gh/wconstab/443/base 2025-09-07T07:36:04.1653382Z * [new branch] gh/wconstab/443/head -> origin/gh/wconstab/443/head 2025-09-07T07:36:04.1654561Z * [new branch] gh/wconstab/443/orig -> origin/gh/wconstab/443/orig 2025-09-07T07:36:04.1656430Z * [new branch] gh/wconstab/444/base -> origin/gh/wconstab/444/base 2025-09-07T07:36:04.1657525Z * [new branch] gh/wconstab/444/head -> origin/gh/wconstab/444/head 2025-09-07T07:36:04.1658860Z * [new branch] gh/wconstab/444/orig -> origin/gh/wconstab/444/orig 2025-09-07T07:36:04.1660545Z * [new branch] gh/wconstab/445/base -> origin/gh/wconstab/445/base 2025-09-07T07:36:04.1661723Z * [new branch] gh/wconstab/445/head -> origin/gh/wconstab/445/head 2025-09-07T07:36:04.1662913Z * [new branch] gh/wconstab/445/orig -> origin/gh/wconstab/445/orig 2025-09-07T07:36:04.1665035Z * [new branch] gh/wconstab/446/base -> origin/gh/wconstab/446/base 2025-09-07T07:36:04.1666347Z * [new branch] gh/wconstab/446/head -> origin/gh/wconstab/446/head 2025-09-07T07:36:04.1667814Z * [new branch] gh/wconstab/446/orig -> origin/gh/wconstab/446/orig 2025-09-07T07:36:04.1669553Z * [new branch] gh/wconstab/447/base -> origin/gh/wconstab/447/base 2025-09-07T07:36:04.1670714Z * [new branch] gh/wconstab/447/head -> origin/gh/wconstab/447/head 2025-09-07T07:36:04.1671926Z * [new branch] gh/wconstab/447/orig -> origin/gh/wconstab/447/orig 2025-09-07T07:36:04.1674222Z * [new branch] gh/weifengpy/27/base -> origin/gh/weifengpy/27/base 2025-09-07T07:36:04.1675417Z * [new branch] gh/weifengpy/27/head -> origin/gh/weifengpy/27/head 2025-09-07T07:36:04.1676590Z * [new branch] gh/weifengpy/27/orig -> origin/gh/weifengpy/27/orig 2025-09-07T07:36:04.1678355Z * [new branch] gh/weifengpy/30/base -> origin/gh/weifengpy/30/base 2025-09-07T07:36:04.1679514Z * [new branch] gh/weifengpy/30/head -> origin/gh/weifengpy/30/head 2025-09-07T07:36:04.1680685Z * [new branch] gh/weifengpy/30/orig -> origin/gh/weifengpy/30/orig 2025-09-07T07:36:04.1683278Z * [new branch] gh/williamwen42/196/base -> origin/gh/williamwen42/196/base 2025-09-07T07:36:04.1684635Z * [new branch] gh/williamwen42/196/head -> origin/gh/williamwen42/196/head 2025-09-07T07:36:04.1685918Z * [new branch] gh/williamwen42/196/orig -> origin/gh/williamwen42/196/orig 2025-09-07T07:36:04.1687660Z * [new branch] gh/williamwen42/250/base -> origin/gh/williamwen42/250/base 2025-09-07T07:36:04.1688882Z * [new branch] gh/williamwen42/250/head -> origin/gh/williamwen42/250/head 2025-09-07T07:36:04.1690110Z * [new branch] gh/williamwen42/250/orig -> origin/gh/williamwen42/250/orig 2025-09-07T07:36:04.1691944Z * [new branch] gh/williamwen42/258/base -> origin/gh/williamwen42/258/base 2025-09-07T07:36:04.1693140Z * [new branch] gh/williamwen42/258/head -> origin/gh/williamwen42/258/head 2025-09-07T07:36:04.1694364Z * [new branch] gh/williamwen42/258/orig -> origin/gh/williamwen42/258/orig 2025-09-07T07:36:04.1696037Z * [new branch] gh/williamwen42/266/base -> origin/gh/williamwen42/266/base 2025-09-07T07:36:04.1697305Z * [new branch] gh/williamwen42/266/head -> origin/gh/williamwen42/266/head 2025-09-07T07:36:04.1698644Z * [new branch] gh/williamwen42/266/orig -> origin/gh/williamwen42/266/orig 2025-09-07T07:36:04.1700378Z * [new branch] gh/williamwen42/267/base -> origin/gh/williamwen42/267/base 2025-09-07T07:36:04.1701633Z * [new branch] gh/williamwen42/267/head -> origin/gh/williamwen42/267/head 2025-09-07T07:36:04.1702794Z * [new branch] gh/williamwen42/267/orig -> origin/gh/williamwen42/267/orig 2025-09-07T07:36:04.1705779Z * [new branch] gh/williamwen42/270/base -> origin/gh/williamwen42/270/base 2025-09-07T07:36:04.1706134Z * [new branch] gh/williamwen42/270/head -> origin/gh/williamwen42/270/head 2025-09-07T07:36:04.1707391Z * [new branch] gh/williamwen42/270/orig -> origin/gh/williamwen42/270/orig 2025-09-07T07:36:04.1708991Z * [new branch] gh/williamwen42/271/base -> origin/gh/williamwen42/271/base 2025-09-07T07:36:04.1710224Z * [new branch] gh/williamwen42/271/head -> origin/gh/williamwen42/271/head 2025-09-07T07:36:04.1711443Z * [new branch] gh/williamwen42/271/orig -> origin/gh/williamwen42/271/orig 2025-09-07T07:36:04.1713164Z * [new branch] gh/williamwen42/272/base -> origin/gh/williamwen42/272/base 2025-09-07T07:36:04.1714372Z * [new branch] gh/williamwen42/272/head -> origin/gh/williamwen42/272/head 2025-09-07T07:36:04.1715601Z * [new branch] gh/williamwen42/272/orig -> origin/gh/williamwen42/272/orig 2025-09-07T07:36:04.1717306Z * [new branch] gh/williamwen42/274/base -> origin/gh/williamwen42/274/base 2025-09-07T07:36:04.1718575Z * [new branch] gh/williamwen42/274/head -> origin/gh/williamwen42/274/head 2025-09-07T07:36:04.1720023Z * [new branch] gh/williamwen42/274/orig -> origin/gh/williamwen42/274/orig 2025-09-07T07:36:04.1721702Z * [new branch] gh/williamwen42/275/base -> origin/gh/williamwen42/275/base 2025-09-07T07:36:04.1722868Z * [new branch] gh/williamwen42/275/head -> origin/gh/williamwen42/275/head 2025-09-07T07:36:04.1724532Z * [new branch] gh/williamwen42/276/base -> origin/gh/williamwen42/276/base 2025-09-07T07:36:04.1725739Z * [new branch] gh/williamwen42/276/head -> origin/gh/williamwen42/276/head 2025-09-07T07:36:04.1726931Z * [new branch] gh/williamwen42/276/orig -> origin/gh/williamwen42/276/orig 2025-09-07T07:36:04.1728712Z * [new branch] gh/williamwen42/277/base -> origin/gh/williamwen42/277/base 2025-09-07T07:36:04.1729899Z * [new branch] gh/williamwen42/277/head -> origin/gh/williamwen42/277/head 2025-09-07T07:36:04.1731087Z * [new branch] gh/williamwen42/277/orig -> origin/gh/williamwen42/277/orig 2025-09-07T07:36:04.1732870Z * [new branch] gh/williamwen42/278/base -> origin/gh/williamwen42/278/base 2025-09-07T07:36:04.1734048Z * [new branch] gh/williamwen42/278/head -> origin/gh/williamwen42/278/head 2025-09-07T07:36:04.1735242Z * [new branch] gh/williamwen42/278/orig -> origin/gh/williamwen42/278/orig 2025-09-07T07:36:04.1736961Z * [new branch] gh/williamwen42/279/base -> origin/gh/williamwen42/279/base 2025-09-07T07:36:04.1738209Z * [new branch] gh/williamwen42/279/head -> origin/gh/williamwen42/279/head 2025-09-07T07:36:04.1739426Z * [new branch] gh/williamwen42/279/orig -> origin/gh/williamwen42/279/orig 2025-09-07T07:36:04.1741348Z * [new branch] gh/williamwen42/280/base -> origin/gh/williamwen42/280/base 2025-09-07T07:36:04.1744987Z * [new branch] gh/williamwen42/280/head -> origin/gh/williamwen42/280/head 2025-09-07T07:36:04.1746737Z * [new branch] gh/williamwen42/280/orig -> origin/gh/williamwen42/280/orig 2025-09-07T07:36:04.1746889Z * [new branch] gh/williamwen42/281/base -> origin/gh/williamwen42/281/base 2025-09-07T07:36:04.1747027Z * [new branch] gh/williamwen42/281/head -> origin/gh/williamwen42/281/head 2025-09-07T07:36:04.1747648Z * [new branch] gh/williamwen42/281/orig -> origin/gh/williamwen42/281/orig 2025-09-07T07:36:04.1749692Z * [new branch] gh/williamwen42/282/base -> origin/gh/williamwen42/282/base 2025-09-07T07:36:04.1750919Z * [new branch] gh/williamwen42/282/head -> origin/gh/williamwen42/282/head 2025-09-07T07:36:04.1752079Z * [new branch] gh/williamwen42/282/orig -> origin/gh/williamwen42/282/orig 2025-09-07T07:36:04.1754082Z * [new branch] gh/williamwen42/283/base -> origin/gh/williamwen42/283/base 2025-09-07T07:36:04.1755643Z * [new branch] gh/williamwen42/283/head -> origin/gh/williamwen42/283/head 2025-09-07T07:36:04.1756783Z * [new branch] gh/williamwen42/283/orig -> origin/gh/williamwen42/283/orig 2025-09-07T07:36:04.1758757Z * [new branch] gh/williamwen42/284/base -> origin/gh/williamwen42/284/base 2025-09-07T07:36:04.1759909Z * [new branch] gh/williamwen42/284/head -> origin/gh/williamwen42/284/head 2025-09-07T07:36:04.1761074Z * [new branch] gh/williamwen42/284/orig -> origin/gh/williamwen42/284/orig 2025-09-07T07:36:04.1762664Z * [new branch] gh/williamwen42/285/base -> origin/gh/williamwen42/285/base 2025-09-07T07:36:04.1763916Z * [new branch] gh/williamwen42/285/head -> origin/gh/williamwen42/285/head 2025-09-07T07:36:04.1765094Z * [new branch] gh/williamwen42/285/orig -> origin/gh/williamwen42/285/orig 2025-09-07T07:36:04.1766717Z * [new branch] gh/williamwen42/286/base -> origin/gh/williamwen42/286/base 2025-09-07T07:36:04.1767876Z * [new branch] gh/williamwen42/286/head -> origin/gh/williamwen42/286/head 2025-09-07T07:36:04.1769048Z * [new branch] gh/williamwen42/286/orig -> origin/gh/williamwen42/286/orig 2025-09-07T07:36:04.1770833Z * [new branch] gh/williamwen42/287/base -> origin/gh/williamwen42/287/base 2025-09-07T07:36:04.1772023Z * [new branch] gh/williamwen42/287/head -> origin/gh/williamwen42/287/head 2025-09-07T07:36:04.1773403Z * [new branch] gh/williamwen42/287/orig -> origin/gh/williamwen42/287/orig 2025-09-07T07:36:04.1775290Z * [new branch] gh/williamwen42/288/base -> origin/gh/williamwen42/288/base 2025-09-07T07:36:04.1776527Z * [new branch] gh/williamwen42/288/head -> origin/gh/williamwen42/288/head 2025-09-07T07:36:04.1777755Z * [new branch] gh/williamwen42/288/orig -> origin/gh/williamwen42/288/orig 2025-09-07T07:36:04.1779537Z * [new branch] gh/williamwen42/289/base -> origin/gh/williamwen42/289/base 2025-09-07T07:36:04.1780719Z * [new branch] gh/williamwen42/289/head -> origin/gh/williamwen42/289/head 2025-09-07T07:36:04.1781896Z * [new branch] gh/williamwen42/289/orig -> origin/gh/williamwen42/289/orig 2025-09-07T07:36:04.1784196Z * [new branch] gh/wychi/1/base -> origin/gh/wychi/1/base 2025-09-07T07:36:04.1785440Z * [new branch] gh/wychi/1/head -> origin/gh/wychi/1/head 2025-09-07T07:36:04.1786662Z * [new branch] gh/wychi/1/orig -> origin/gh/wychi/1/orig 2025-09-07T07:36:04.1788772Z * [new branch] gh/xmfan/169/base -> origin/gh/xmfan/169/base 2025-09-07T07:36:04.1790038Z * [new branch] gh/xmfan/169/head -> origin/gh/xmfan/169/head 2025-09-07T07:36:04.1791771Z * [new branch] gh/xmfan/170/base -> origin/gh/xmfan/170/base 2025-09-07T07:36:04.1792960Z * [new branch] gh/xmfan/170/head -> origin/gh/xmfan/170/head 2025-09-07T07:36:04.1794753Z * [new branch] gh/xmfan/18/base -> origin/gh/xmfan/18/base 2025-09-07T07:36:04.1795922Z * [new branch] gh/xmfan/18/head -> origin/gh/xmfan/18/head 2025-09-07T07:36:04.1797571Z * [new branch] gh/xmfan/229/base -> origin/gh/xmfan/229/base 2025-09-07T07:36:04.1798749Z * [new branch] gh/xmfan/229/head -> origin/gh/xmfan/229/head 2025-09-07T07:36:04.1799904Z * [new branch] gh/xmfan/229/orig -> origin/gh/xmfan/229/orig 2025-09-07T07:36:04.1801580Z * [new branch] gh/xmfan/237/base -> origin/gh/xmfan/237/base 2025-09-07T07:36:04.1802861Z * [new branch] gh/xmfan/237/head -> origin/gh/xmfan/237/head 2025-09-07T07:36:04.1804364Z * [new branch] gh/xmfan/237/orig -> origin/gh/xmfan/237/orig 2025-09-07T07:36:04.1806111Z * [new branch] gh/xmfan/244/base -> origin/gh/xmfan/244/base 2025-09-07T07:36:04.1807270Z * [new branch] gh/xmfan/244/head -> origin/gh/xmfan/244/head 2025-09-07T07:36:04.1808443Z * [new branch] gh/xmfan/244/orig -> origin/gh/xmfan/244/orig 2025-09-07T07:36:04.1810125Z * [new branch] gh/xmfan/246/base -> origin/gh/xmfan/246/base 2025-09-07T07:36:04.1811327Z * [new branch] gh/xmfan/246/head -> origin/gh/xmfan/246/head 2025-09-07T07:36:04.1812491Z * [new branch] gh/xmfan/246/orig -> origin/gh/xmfan/246/orig 2025-09-07T07:36:04.1814202Z * [new branch] gh/xmfan/253/base -> origin/gh/xmfan/253/base 2025-09-07T07:36:04.1815405Z * [new branch] gh/xmfan/253/head -> origin/gh/xmfan/253/head 2025-09-07T07:36:04.1816591Z * [new branch] gh/xmfan/253/orig -> origin/gh/xmfan/253/orig 2025-09-07T07:36:04.1818338Z * [new branch] gh/xmfan/254/base -> origin/gh/xmfan/254/base 2025-09-07T07:36:04.1819518Z * [new branch] gh/xmfan/254/head -> origin/gh/xmfan/254/head 2025-09-07T07:36:04.1820706Z * [new branch] gh/xmfan/254/orig -> origin/gh/xmfan/254/orig 2025-09-07T07:36:04.1822510Z * [new branch] gh/xmfan/260/base -> origin/gh/xmfan/260/base 2025-09-07T07:36:04.1823708Z * [new branch] gh/xmfan/260/head -> origin/gh/xmfan/260/head 2025-09-07T07:36:04.1824846Z * [new branch] gh/xmfan/260/orig -> origin/gh/xmfan/260/orig 2025-09-07T07:36:04.1826543Z * [new branch] gh/xmfan/262/base -> origin/gh/xmfan/262/base 2025-09-07T07:36:04.1827879Z * [new branch] gh/xmfan/262/head -> origin/gh/xmfan/262/head 2025-09-07T07:36:04.1829073Z * [new branch] gh/xmfan/262/orig -> origin/gh/xmfan/262/orig 2025-09-07T07:36:04.1830780Z * [new branch] gh/xmfan/263/base -> origin/gh/xmfan/263/base 2025-09-07T07:36:04.1831980Z * [new branch] gh/xmfan/263/head -> origin/gh/xmfan/263/head 2025-09-07T07:36:04.1833211Z * [new branch] gh/xmfan/263/orig -> origin/gh/xmfan/263/orig 2025-09-07T07:36:04.1834883Z * [new branch] gh/xmfan/264/base -> origin/gh/xmfan/264/base 2025-09-07T07:36:04.1836085Z * [new branch] gh/xmfan/264/head -> origin/gh/xmfan/264/head 2025-09-07T07:36:04.1837272Z * [new branch] gh/xmfan/264/orig -> origin/gh/xmfan/264/orig 2025-09-07T07:36:04.1838998Z * [new branch] gh/xmfan/274/base -> origin/gh/xmfan/274/base 2025-09-07T07:36:04.1840162Z * [new branch] gh/xmfan/274/head -> origin/gh/xmfan/274/head 2025-09-07T07:36:04.1841404Z * [new branch] gh/xmfan/274/orig -> origin/gh/xmfan/274/orig 2025-09-07T07:36:04.1843142Z * [new branch] gh/xmfan/276/base -> origin/gh/xmfan/276/base 2025-09-07T07:36:04.1844760Z * [new branch] gh/xmfan/276/head -> origin/gh/xmfan/276/head 2025-09-07T07:36:04.1846009Z * [new branch] gh/xmfan/276/orig -> origin/gh/xmfan/276/orig 2025-09-07T07:36:04.1847674Z * [new branch] gh/xmfan/277/base -> origin/gh/xmfan/277/base 2025-09-07T07:36:04.1848861Z * [new branch] gh/xmfan/277/head -> origin/gh/xmfan/277/head 2025-09-07T07:36:04.1850023Z * [new branch] gh/xmfan/277/orig -> origin/gh/xmfan/277/orig 2025-09-07T07:36:04.1851720Z * [new branch] gh/xmfan/278/base -> origin/gh/xmfan/278/base 2025-09-07T07:36:04.1852884Z * [new branch] gh/xmfan/278/head -> origin/gh/xmfan/278/head 2025-09-07T07:36:04.1854186Z * [new branch] gh/xmfan/278/orig -> origin/gh/xmfan/278/orig 2025-09-07T07:36:04.1856221Z * [new branch] gh/xmfan/279/base -> origin/gh/xmfan/279/base 2025-09-07T07:36:04.1857407Z * [new branch] gh/xmfan/279/head -> origin/gh/xmfan/279/head 2025-09-07T07:36:04.1858661Z * [new branch] gh/xmfan/279/orig -> origin/gh/xmfan/279/orig 2025-09-07T07:36:04.1860483Z * [new branch] gh/xmfan/280/base -> origin/gh/xmfan/280/base 2025-09-07T07:36:04.1861635Z * [new branch] gh/xmfan/280/head -> origin/gh/xmfan/280/head 2025-09-07T07:36:04.1862941Z * [new branch] gh/xmfan/280/orig -> origin/gh/xmfan/280/orig 2025-09-07T07:36:04.1864675Z * [new branch] gh/xmfan/281/base -> origin/gh/xmfan/281/base 2025-09-07T07:36:04.1865870Z * [new branch] gh/xmfan/281/head -> origin/gh/xmfan/281/head 2025-09-07T07:36:04.1867042Z * [new branch] gh/xmfan/281/orig -> origin/gh/xmfan/281/orig 2025-09-07T07:36:04.1868757Z * [new branch] gh/xmfan/282/base -> origin/gh/xmfan/282/base 2025-09-07T07:36:04.1870015Z * [new branch] gh/xmfan/282/head -> origin/gh/xmfan/282/head 2025-09-07T07:36:04.1871765Z * [new branch] gh/xmfan/283/base -> origin/gh/xmfan/283/base 2025-09-07T07:36:04.1873029Z * [new branch] gh/xmfan/283/head -> origin/gh/xmfan/283/head 2025-09-07T07:36:04.1874272Z * [new branch] gh/xmfan/283/orig -> origin/gh/xmfan/283/orig 2025-09-07T07:36:04.1876436Z * [new branch] gh/xuanzhang816/14/base -> origin/gh/xuanzhang816/14/base 2025-09-07T07:36:04.1880494Z * [new branch] gh/xuanzhang816/14/head -> origin/gh/xuanzhang816/14/head 2025-09-07T07:36:04.1881739Z * [new branch] gh/xuanzhang816/14/orig -> origin/gh/xuanzhang816/14/orig 2025-09-07T07:36:04.1883423Z * [new branch] gh/xuanzhang816/19/base -> origin/gh/xuanzhang816/19/base 2025-09-07T07:36:04.1884589Z * [new branch] gh/xuanzhang816/19/head -> origin/gh/xuanzhang816/19/head 2025-09-07T07:36:04.1885826Z * [new branch] gh/xuanzhang816/19/orig -> origin/gh/xuanzhang816/19/orig 2025-09-07T07:36:04.1887537Z * [new branch] gh/xuanzhang816/22/base -> origin/gh/xuanzhang816/22/base 2025-09-07T07:36:04.1888732Z * [new branch] gh/xuanzhang816/22/head -> origin/gh/xuanzhang816/22/head 2025-09-07T07:36:04.1889915Z * [new branch] gh/xuanzhang816/22/orig -> origin/gh/xuanzhang816/22/orig 2025-09-07T07:36:04.1891701Z * [new branch] gh/xuanzhang816/23/base -> origin/gh/xuanzhang816/23/base 2025-09-07T07:36:04.1892909Z * [new branch] gh/xuanzhang816/23/head -> origin/gh/xuanzhang816/23/head 2025-09-07T07:36:04.1894089Z * [new branch] gh/xuanzhang816/23/orig -> origin/gh/xuanzhang816/23/orig 2025-09-07T07:36:04.1895756Z * [new branch] gh/xuanzhang816/24/base -> origin/gh/xuanzhang816/24/base 2025-09-07T07:36:04.1896933Z * [new branch] gh/xuanzhang816/24/head -> origin/gh/xuanzhang816/24/head 2025-09-07T07:36:04.1898565Z * [new branch] gh/xuanzhang816/24/orig -> origin/gh/xuanzhang816/24/orig 2025-09-07T07:36:04.1900273Z * [new branch] gh/xuanzhang816/25/base -> origin/gh/xuanzhang816/25/base 2025-09-07T07:36:04.1901648Z * [new branch] gh/xuanzhang816/25/head -> origin/gh/xuanzhang816/25/head 2025-09-07T07:36:04.1902826Z * [new branch] gh/xuanzhang816/25/orig -> origin/gh/xuanzhang816/25/orig 2025-09-07T07:36:04.1905138Z * [new branch] gh/xuanzhang816/26/base -> origin/gh/xuanzhang816/26/base 2025-09-07T07:36:04.1906337Z * [new branch] gh/xuanzhang816/26/head -> origin/gh/xuanzhang816/26/head 2025-09-07T07:36:04.1907647Z * [new branch] gh/xuanzhang816/26/orig -> origin/gh/xuanzhang816/26/orig 2025-09-07T07:36:04.1909853Z * [new branch] gh/yanbing-j/11/base -> origin/gh/yanbing-j/11/base 2025-09-07T07:36:04.1911048Z * [new branch] gh/yanbing-j/11/head -> origin/gh/yanbing-j/11/head 2025-09-07T07:36:04.1912254Z * [new branch] gh/yanbing-j/11/orig -> origin/gh/yanbing-j/11/orig 2025-09-07T07:36:04.1914053Z * [new branch] gh/yanbing-j/12/base -> origin/gh/yanbing-j/12/base 2025-09-07T07:36:04.1915666Z * [new branch] gh/yanbing-j/12/head -> origin/gh/yanbing-j/12/head 2025-09-07T07:36:04.1916903Z * [new branch] gh/yanbing-j/12/orig -> origin/gh/yanbing-j/12/orig 2025-09-07T07:36:04.1918630Z * [new branch] gh/yanbing-j/13/base -> origin/gh/yanbing-j/13/base 2025-09-07T07:36:04.1919828Z * [new branch] gh/yanbing-j/13/head -> origin/gh/yanbing-j/13/head 2025-09-07T07:36:04.1921017Z * [new branch] gh/yanbing-j/13/orig -> origin/gh/yanbing-j/13/orig 2025-09-07T07:36:04.1922778Z * [new branch] gh/yanbing-j/14/base -> origin/gh/yanbing-j/14/base 2025-09-07T07:36:04.1924013Z * [new branch] gh/yanbing-j/14/head -> origin/gh/yanbing-j/14/head 2025-09-07T07:36:04.1925219Z * [new branch] gh/yanbing-j/14/orig -> origin/gh/yanbing-j/14/orig 2025-09-07T07:36:04.1926881Z * [new branch] gh/yanbing-j/15/base -> origin/gh/yanbing-j/15/base 2025-09-07T07:36:04.1928074Z * [new branch] gh/yanbing-j/15/head -> origin/gh/yanbing-j/15/head 2025-09-07T07:36:04.1929271Z * [new branch] gh/yanbing-j/15/orig -> origin/gh/yanbing-j/15/orig 2025-09-07T07:36:04.1930929Z * [new branch] gh/yanbing-j/18/base -> origin/gh/yanbing-j/18/base 2025-09-07T07:36:04.1932119Z * [new branch] gh/yanbing-j/18/head -> origin/gh/yanbing-j/18/head 2025-09-07T07:36:04.1933304Z * [new branch] gh/yanbing-j/18/orig -> origin/gh/yanbing-j/18/orig 2025-09-07T07:36:04.1935022Z * [new branch] gh/yanbing-j/19/base -> origin/gh/yanbing-j/19/base 2025-09-07T07:36:04.1936231Z * [new branch] gh/yanbing-j/19/head -> origin/gh/yanbing-j/19/head 2025-09-07T07:36:04.1937530Z * [new branch] gh/yanbing-j/19/orig -> origin/gh/yanbing-j/19/orig 2025-09-07T07:36:04.1939407Z * [new branch] gh/yanbing-j/20/base -> origin/gh/yanbing-j/20/base 2025-09-07T07:36:04.1940607Z * [new branch] gh/yanbing-j/20/head -> origin/gh/yanbing-j/20/head 2025-09-07T07:36:04.1941847Z * [new branch] gh/yanbing-j/20/orig -> origin/gh/yanbing-j/20/orig 2025-09-07T07:36:04.1943576Z * [new branch] gh/yanbing-j/21/base -> origin/gh/yanbing-j/21/base 2025-09-07T07:36:04.1944797Z * [new branch] gh/yanbing-j/21/head -> origin/gh/yanbing-j/21/head 2025-09-07T07:36:04.1946448Z * [new branch] gh/yanbing-j/22/base -> origin/gh/yanbing-j/22/base 2025-09-07T07:36:04.1947649Z * [new branch] gh/yanbing-j/22/head -> origin/gh/yanbing-j/22/head 2025-09-07T07:36:04.1948854Z * [new branch] gh/yanbing-j/22/orig -> origin/gh/yanbing-j/22/orig 2025-09-07T07:36:04.1950619Z * [new branch] gh/yanbing-j/23/base -> origin/gh/yanbing-j/23/base 2025-09-07T07:36:04.1951750Z * [new branch] gh/yanbing-j/23/head -> origin/gh/yanbing-j/23/head 2025-09-07T07:36:04.1952917Z * [new branch] gh/yanbing-j/23/orig -> origin/gh/yanbing-j/23/orig 2025-09-07T07:36:04.1954629Z * [new branch] gh/yanbing-j/24/base -> origin/gh/yanbing-j/24/base 2025-09-07T07:36:04.1955818Z * [new branch] gh/yanbing-j/24/head -> origin/gh/yanbing-j/24/head 2025-09-07T07:36:04.1957041Z * [new branch] gh/yanbing-j/24/orig -> origin/gh/yanbing-j/24/orig 2025-09-07T07:36:04.1958721Z * [new branch] gh/yanbing-j/25/base -> origin/gh/yanbing-j/25/base 2025-09-07T07:36:04.1959901Z * [new branch] gh/yanbing-j/25/head -> origin/gh/yanbing-j/25/head 2025-09-07T07:36:04.1961079Z * [new branch] gh/yanbing-j/25/orig -> origin/gh/yanbing-j/25/orig 2025-09-07T07:36:04.1962766Z * [new branch] gh/yanbing-j/26/base -> origin/gh/yanbing-j/26/base 2025-09-07T07:36:04.1963924Z * [new branch] gh/yanbing-j/26/head -> origin/gh/yanbing-j/26/head 2025-09-07T07:36:04.1965636Z * [new branch] gh/yanbing-j/26/orig -> origin/gh/yanbing-j/26/orig 2025-09-07T07:36:04.1967350Z * [new branch] gh/yanbing-j/36/base -> origin/gh/yanbing-j/36/base 2025-09-07T07:36:04.1968524Z * [new branch] gh/yanbing-j/36/head -> origin/gh/yanbing-j/36/head 2025-09-07T07:36:04.1969718Z * [new branch] gh/yanbing-j/36/orig -> origin/gh/yanbing-j/36/orig 2025-09-07T07:36:04.1971438Z * [new branch] gh/yanbing-j/37/base -> origin/gh/yanbing-j/37/base 2025-09-07T07:36:04.1972776Z * [new branch] gh/yanbing-j/37/head -> origin/gh/yanbing-j/37/head 2025-09-07T07:36:04.1973945Z * [new branch] gh/yanbing-j/37/orig -> origin/gh/yanbing-j/37/orig 2025-09-07T07:36:04.1976151Z * [new branch] gh/yangw-dev/12/base -> origin/gh/yangw-dev/12/base 2025-09-07T07:36:04.1977338Z * [new branch] gh/yangw-dev/12/head -> origin/gh/yangw-dev/12/head 2025-09-07T07:36:04.1978620Z * [new branch] gh/yangw-dev/12/orig -> origin/gh/yangw-dev/12/orig 2025-09-07T07:36:04.1980327Z * [new branch] gh/yangw-dev/13/base -> origin/gh/yangw-dev/13/base 2025-09-07T07:36:04.1981545Z * [new branch] gh/yangw-dev/13/head -> origin/gh/yangw-dev/13/head 2025-09-07T07:36:04.1982715Z * [new branch] gh/yangw-dev/13/orig -> origin/gh/yangw-dev/13/orig 2025-09-07T07:36:04.1984840Z * [new branch] gh/yangw-dev/14/base -> origin/gh/yangw-dev/14/base 2025-09-07T07:36:04.1986028Z * [new branch] gh/yangw-dev/14/head -> origin/gh/yangw-dev/14/head 2025-09-07T07:36:04.1987249Z * [new branch] gh/yangw-dev/14/orig -> origin/gh/yangw-dev/14/orig 2025-09-07T07:36:04.1988975Z * [new branch] gh/yangw-dev/15/base -> origin/gh/yangw-dev/15/base 2025-09-07T07:36:04.1990164Z * [new branch] gh/yangw-dev/15/head -> origin/gh/yangw-dev/15/head 2025-09-07T07:36:04.1991365Z * [new branch] gh/yangw-dev/15/orig -> origin/gh/yangw-dev/15/orig 2025-09-07T07:36:04.1993117Z * [new branch] gh/yangw-dev/16/base -> origin/gh/yangw-dev/16/base 2025-09-07T07:36:04.1994297Z * [new branch] gh/yangw-dev/16/head -> origin/gh/yangw-dev/16/head 2025-09-07T07:36:04.1995443Z * [new branch] gh/yangw-dev/16/orig -> origin/gh/yangw-dev/16/orig 2025-09-07T07:36:04.1997141Z * [new branch] gh/yangw-dev/17/base -> origin/gh/yangw-dev/17/base 2025-09-07T07:36:04.1998335Z * [new branch] gh/yangw-dev/17/head -> origin/gh/yangw-dev/17/head 2025-09-07T07:36:04.1999469Z * [new branch] gh/yangw-dev/17/orig -> origin/gh/yangw-dev/17/orig 2025-09-07T07:36:04.2001130Z * [new branch] gh/yangw-dev/18/base -> origin/gh/yangw-dev/18/base 2025-09-07T07:36:04.2002341Z * [new branch] gh/yangw-dev/18/head -> origin/gh/yangw-dev/18/head 2025-09-07T07:36:04.2003736Z * [new branch] gh/yangw-dev/18/orig -> origin/gh/yangw-dev/18/orig 2025-09-07T07:36:04.2005551Z * [new branch] gh/yangw-dev/19/base -> origin/gh/yangw-dev/19/base 2025-09-07T07:36:04.2007055Z * [new branch] gh/yangw-dev/19/head -> origin/gh/yangw-dev/19/head 2025-09-07T07:36:04.2007949Z * [new branch] gh/yangw-dev/19/orig -> origin/gh/yangw-dev/19/orig 2025-09-07T07:36:04.2009729Z * [new branch] gh/yangw-dev/20/base -> origin/gh/yangw-dev/20/base 2025-09-07T07:36:04.2010904Z * [new branch] gh/yangw-dev/20/head -> origin/gh/yangw-dev/20/head 2025-09-07T07:36:04.2012489Z * [new branch] gh/yangw-dev/20/orig -> origin/gh/yangw-dev/20/orig 2025-09-07T07:36:04.2014240Z * [new branch] gh/yangw-dev/21/base -> origin/gh/yangw-dev/21/base 2025-09-07T07:36:04.2015445Z * [new branch] gh/yangw-dev/21/head -> origin/gh/yangw-dev/21/head 2025-09-07T07:36:04.2016652Z * [new branch] gh/yangw-dev/21/orig -> origin/gh/yangw-dev/21/orig 2025-09-07T07:36:04.2018471Z * [new branch] gh/yangw-dev/22/base -> origin/gh/yangw-dev/22/base 2025-09-07T07:36:04.2019753Z * [new branch] gh/yangw-dev/22/head -> origin/gh/yangw-dev/22/head 2025-09-07T07:36:04.2020916Z * [new branch] gh/yangw-dev/22/orig -> origin/gh/yangw-dev/22/orig 2025-09-07T07:36:04.2022549Z * [new branch] gh/yangw-dev/23/base -> origin/gh/yangw-dev/23/base 2025-09-07T07:36:04.2023714Z * [new branch] gh/yangw-dev/23/head -> origin/gh/yangw-dev/23/head 2025-09-07T07:36:04.2024933Z * [new branch] gh/yangw-dev/23/orig -> origin/gh/yangw-dev/23/orig 2025-09-07T07:36:04.2027057Z * [new branch] gh/yangw-dev/24/base -> origin/gh/yangw-dev/24/base 2025-09-07T07:36:04.2028256Z * [new branch] gh/yangw-dev/24/head -> origin/gh/yangw-dev/24/head 2025-09-07T07:36:04.2029417Z * [new branch] gh/yangw-dev/24/orig -> origin/gh/yangw-dev/24/orig 2025-09-07T07:36:04.2031137Z * [new branch] gh/yangw-dev/25/base -> origin/gh/yangw-dev/25/base 2025-09-07T07:36:04.2032298Z * [new branch] gh/yangw-dev/25/head -> origin/gh/yangw-dev/25/head 2025-09-07T07:36:04.2033467Z * [new branch] gh/yangw-dev/25/orig -> origin/gh/yangw-dev/25/orig 2025-09-07T07:36:04.2035177Z * [new branch] gh/yangw-dev/26/base -> origin/gh/yangw-dev/26/base 2025-09-07T07:36:04.2036356Z * [new branch] gh/yangw-dev/26/head -> origin/gh/yangw-dev/26/head 2025-09-07T07:36:04.2037526Z * [new branch] gh/yangw-dev/26/orig -> origin/gh/yangw-dev/26/orig 2025-09-07T07:36:04.2039331Z * [new branch] gh/yangw-dev/27/base -> origin/gh/yangw-dev/27/base 2025-09-07T07:36:04.2040546Z * [new branch] gh/yangw-dev/27/head -> origin/gh/yangw-dev/27/head 2025-09-07T07:36:04.2041687Z * [new branch] gh/yangw-dev/27/orig -> origin/gh/yangw-dev/27/orig 2025-09-07T07:36:04.2043865Z * [new branch] gh/ydwu4/233/base -> origin/gh/ydwu4/233/base 2025-09-07T07:36:04.2045252Z * [new branch] gh/ydwu4/233/head -> origin/gh/ydwu4/233/head 2025-09-07T07:36:04.2046445Z * [new branch] gh/ydwu4/233/orig -> origin/gh/ydwu4/233/orig 2025-09-07T07:36:04.2048307Z * [new branch] gh/ydwu4/246/base -> origin/gh/ydwu4/246/base 2025-09-07T07:36:04.2049898Z * [new branch] gh/ydwu4/246/head -> origin/gh/ydwu4/246/head 2025-09-07T07:36:04.2051156Z * [new branch] gh/ydwu4/246/orig -> origin/gh/ydwu4/246/orig 2025-09-07T07:36:04.2053028Z * [new branch] gh/ydwu4/253/base -> origin/gh/ydwu4/253/base 2025-09-07T07:36:04.2054279Z * [new branch] gh/ydwu4/253/head -> origin/gh/ydwu4/253/head 2025-09-07T07:36:04.2055495Z * [new branch] gh/ydwu4/253/orig -> origin/gh/ydwu4/253/orig 2025-09-07T07:36:04.2057214Z * [new branch] gh/ydwu4/255/base -> origin/gh/ydwu4/255/base 2025-09-07T07:36:04.2058573Z * [new branch] gh/ydwu4/255/head -> origin/gh/ydwu4/255/head 2025-09-07T07:36:04.2059756Z * [new branch] gh/ydwu4/255/orig -> origin/gh/ydwu4/255/orig 2025-09-07T07:36:04.2061900Z * [new branch] gh/ydwu4/259/base -> origin/gh/ydwu4/259/base 2025-09-07T07:36:04.2063104Z * [new branch] gh/ydwu4/259/head -> origin/gh/ydwu4/259/head 2025-09-07T07:36:04.2064292Z * [new branch] gh/ydwu4/259/orig -> origin/gh/ydwu4/259/orig 2025-09-07T07:36:04.2066038Z * [new branch] gh/ydwu4/262/base -> origin/gh/ydwu4/262/base 2025-09-07T07:36:04.2067265Z * [new branch] gh/ydwu4/262/head -> origin/gh/ydwu4/262/head 2025-09-07T07:36:04.2068474Z * [new branch] gh/ydwu4/262/orig -> origin/gh/ydwu4/262/orig 2025-09-07T07:36:04.2070188Z * [new branch] gh/ydwu4/263/base -> origin/gh/ydwu4/263/base 2025-09-07T07:36:04.2071365Z * [new branch] gh/ydwu4/263/head -> origin/gh/ydwu4/263/head 2025-09-07T07:36:04.2072571Z * [new branch] gh/ydwu4/263/orig -> origin/gh/ydwu4/263/orig 2025-09-07T07:36:04.2074378Z * [new branch] gh/ydwu4/269/base -> origin/gh/ydwu4/269/base 2025-09-07T07:36:04.2075547Z * [new branch] gh/ydwu4/269/head -> origin/gh/ydwu4/269/head 2025-09-07T07:36:04.2076737Z * [new branch] gh/ydwu4/269/orig -> origin/gh/ydwu4/269/orig 2025-09-07T07:36:04.2078481Z * [new branch] gh/ydwu4/270/base -> origin/gh/ydwu4/270/base 2025-09-07T07:36:04.2079717Z * [new branch] gh/ydwu4/270/head -> origin/gh/ydwu4/270/head 2025-09-07T07:36:04.2081487Z * [new branch] gh/ydwu4/270/orig -> origin/gh/ydwu4/270/orig 2025-09-07T07:36:04.2083209Z * [new branch] gh/ydwu4/272/base -> origin/gh/ydwu4/272/base 2025-09-07T07:36:04.2084488Z * [new branch] gh/ydwu4/272/head -> origin/gh/ydwu4/272/head 2025-09-07T07:36:04.2085756Z * [new branch] gh/ydwu4/272/orig -> origin/gh/ydwu4/272/orig 2025-09-07T07:36:04.2087322Z * [new branch] gh/ydwu4/275/base -> origin/gh/ydwu4/275/base 2025-09-07T07:36:04.2088499Z * [new branch] gh/ydwu4/275/head -> origin/gh/ydwu4/275/head 2025-09-07T07:36:04.2089663Z * [new branch] gh/ydwu4/275/orig -> origin/gh/ydwu4/275/orig 2025-09-07T07:36:04.2091251Z * [new branch] gh/ydwu4/276/base -> origin/gh/ydwu4/276/base 2025-09-07T07:36:04.2092440Z * [new branch] gh/ydwu4/276/head -> origin/gh/ydwu4/276/head 2025-09-07T07:36:04.2093787Z * [new branch] gh/ydwu4/276/orig -> origin/gh/ydwu4/276/orig 2025-09-07T07:36:04.2095786Z * [new branch] gh/ydwu4/279/base -> origin/gh/ydwu4/279/base 2025-09-07T07:36:04.2097050Z * [new branch] gh/ydwu4/279/head -> origin/gh/ydwu4/279/head 2025-09-07T07:36:04.2098379Z * [new branch] gh/ydwu4/279/orig -> origin/gh/ydwu4/279/orig 2025-09-07T07:36:04.2100541Z * [new branch] gh/ydwu4/283/base -> origin/gh/ydwu4/283/base 2025-09-07T07:36:04.2101814Z * [new branch] gh/ydwu4/283/head -> origin/gh/ydwu4/283/head 2025-09-07T07:36:04.2102988Z * [new branch] gh/ydwu4/283/orig -> origin/gh/ydwu4/283/orig 2025-09-07T07:36:04.2108824Z * [new branch] gh/ydwu4/289/base -> origin/gh/ydwu4/289/base 2025-09-07T07:36:04.2110041Z * [new branch] gh/ydwu4/289/head -> origin/gh/ydwu4/289/head 2025-09-07T07:36:04.2111219Z * [new branch] gh/ydwu4/289/orig -> origin/gh/ydwu4/289/orig 2025-09-07T07:36:04.2113178Z * [new branch] gh/ydwu4/290/base -> origin/gh/ydwu4/290/base 2025-09-07T07:36:04.2114290Z * [new branch] gh/ydwu4/290/head -> origin/gh/ydwu4/290/head 2025-09-07T07:36:04.2115542Z * [new branch] gh/ydwu4/290/orig -> origin/gh/ydwu4/290/orig 2025-09-07T07:36:04.2118108Z * [new branch] gh/ydwu4/291/base -> origin/gh/ydwu4/291/base 2025-09-07T07:36:04.2119353Z * [new branch] gh/ydwu4/291/head -> origin/gh/ydwu4/291/head 2025-09-07T07:36:04.2121069Z * [new branch] gh/ydwu4/291/orig -> origin/gh/ydwu4/291/orig 2025-09-07T07:36:04.2123008Z * [new branch] gh/ydwu4/292/base -> origin/gh/ydwu4/292/base 2025-09-07T07:36:04.2124258Z * [new branch] gh/ydwu4/292/head -> origin/gh/ydwu4/292/head 2025-09-07T07:36:04.2125453Z * [new branch] gh/ydwu4/292/orig -> origin/gh/ydwu4/292/orig 2025-09-07T07:36:04.2127192Z * [new branch] gh/ydwu4/293/base -> origin/gh/ydwu4/293/base 2025-09-07T07:36:04.2128440Z * [new branch] gh/ydwu4/293/head -> origin/gh/ydwu4/293/head 2025-09-07T07:36:04.2129645Z * [new branch] gh/ydwu4/293/orig -> origin/gh/ydwu4/293/orig 2025-09-07T07:36:04.2131472Z * [new branch] gh/ydwu4/294/base -> origin/gh/ydwu4/294/base 2025-09-07T07:36:04.2132633Z * [new branch] gh/ydwu4/294/head -> origin/gh/ydwu4/294/head 2025-09-07T07:36:04.2133891Z * [new branch] gh/ydwu4/294/orig -> origin/gh/ydwu4/294/orig 2025-09-07T07:36:04.2135662Z * [new branch] gh/ydwu4/295/base -> origin/gh/ydwu4/295/base 2025-09-07T07:36:04.2136866Z * [new branch] gh/ydwu4/295/head -> origin/gh/ydwu4/295/head 2025-09-07T07:36:04.2138121Z * [new branch] gh/ydwu4/295/orig -> origin/gh/ydwu4/295/orig 2025-09-07T07:36:04.2139939Z * [new branch] gh/ydwu4/296/base -> origin/gh/ydwu4/296/base 2025-09-07T07:36:04.2141083Z * [new branch] gh/ydwu4/296/head -> origin/gh/ydwu4/296/head 2025-09-07T07:36:04.2142320Z * [new branch] gh/ydwu4/296/orig -> origin/gh/ydwu4/296/orig 2025-09-07T07:36:04.2144727Z * [new branch] gh/ydwu4/300/base -> origin/gh/ydwu4/300/base 2025-09-07T07:36:04.2146320Z * [new branch] gh/ydwu4/300/head -> origin/gh/ydwu4/300/head 2025-09-07T07:36:04.2147615Z * [new branch] gh/ydwu4/300/orig -> origin/gh/ydwu4/300/orig 2025-09-07T07:36:04.2150069Z * [new branch] gh/ydwu4/301/base -> origin/gh/ydwu4/301/base 2025-09-07T07:36:04.2151225Z * [new branch] gh/ydwu4/301/head -> origin/gh/ydwu4/301/head 2025-09-07T07:36:04.2152467Z * [new branch] gh/ydwu4/301/orig -> origin/gh/ydwu4/301/orig 2025-09-07T07:36:04.2154189Z * [new branch] gh/ydwu4/302/base -> origin/gh/ydwu4/302/base 2025-09-07T07:36:04.2155381Z * [new branch] gh/ydwu4/302/head -> origin/gh/ydwu4/302/head 2025-09-07T07:36:04.2156580Z * [new branch] gh/ydwu4/302/orig -> origin/gh/ydwu4/302/orig 2025-09-07T07:36:04.2158181Z * [new branch] gh/ydwu4/303/base -> origin/gh/ydwu4/303/base 2025-09-07T07:36:04.2159474Z * [new branch] gh/ydwu4/303/head -> origin/gh/ydwu4/303/head 2025-09-07T07:36:04.2160732Z * [new branch] gh/ydwu4/303/orig -> origin/gh/ydwu4/303/orig 2025-09-07T07:36:04.2162789Z * [new branch] gh/ydwu4/304/base -> origin/gh/ydwu4/304/base 2025-09-07T07:36:04.2164030Z * [new branch] gh/ydwu4/304/head -> origin/gh/ydwu4/304/head 2025-09-07T07:36:04.2165237Z * [new branch] gh/ydwu4/304/orig -> origin/gh/ydwu4/304/orig 2025-09-07T07:36:04.2167139Z * [new branch] gh/ydwu4/305/base -> origin/gh/ydwu4/305/base 2025-09-07T07:36:04.2168411Z * [new branch] gh/ydwu4/305/head -> origin/gh/ydwu4/305/head 2025-09-07T07:36:04.2169664Z * [new branch] gh/ydwu4/305/orig -> origin/gh/ydwu4/305/orig 2025-09-07T07:36:04.2171576Z * [new branch] gh/ydwu4/306/base -> origin/gh/ydwu4/306/base 2025-09-07T07:36:04.2172812Z * [new branch] gh/ydwu4/306/head -> origin/gh/ydwu4/306/head 2025-09-07T07:36:04.2174016Z * [new branch] gh/ydwu4/306/orig -> origin/gh/ydwu4/306/orig 2025-09-07T07:36:04.2175693Z * [new branch] gh/ydwu4/307/base -> origin/gh/ydwu4/307/base 2025-09-07T07:36:04.2176831Z * [new branch] gh/ydwu4/307/head -> origin/gh/ydwu4/307/head 2025-09-07T07:36:04.2178058Z * [new branch] gh/ydwu4/307/orig -> origin/gh/ydwu4/307/orig 2025-09-07T07:36:04.2179858Z * [new branch] gh/ydwu4/308/base -> origin/gh/ydwu4/308/base 2025-09-07T07:36:04.2181132Z * [new branch] gh/ydwu4/308/head -> origin/gh/ydwu4/308/head 2025-09-07T07:36:04.2182285Z * [new branch] gh/ydwu4/308/orig -> origin/gh/ydwu4/308/orig 2025-09-07T07:36:04.2183999Z * [new branch] gh/ydwu4/309/base -> origin/gh/ydwu4/309/base 2025-09-07T07:36:04.2185144Z * [new branch] gh/ydwu4/309/head -> origin/gh/ydwu4/309/head 2025-09-07T07:36:04.2186439Z * [new branch] gh/ydwu4/309/orig -> origin/gh/ydwu4/309/orig 2025-09-07T07:36:04.2188286Z * [new branch] gh/ydwu4/310/base -> origin/gh/ydwu4/310/base 2025-09-07T07:36:04.2189669Z * [new branch] gh/ydwu4/310/head -> origin/gh/ydwu4/310/head 2025-09-07T07:36:04.2190842Z * [new branch] gh/ydwu4/310/orig -> origin/gh/ydwu4/310/orig 2025-09-07T07:36:04.2192566Z * [new branch] gh/ydwu4/311/base -> origin/gh/ydwu4/311/base 2025-09-07T07:36:04.2193738Z * [new branch] gh/ydwu4/311/head -> origin/gh/ydwu4/311/head 2025-09-07T07:36:04.2195051Z * [new branch] gh/ydwu4/311/orig -> origin/gh/ydwu4/311/orig 2025-09-07T07:36:04.2196782Z * [new branch] gh/ydwu4/312/base -> origin/gh/ydwu4/312/base 2025-09-07T07:36:04.2198035Z * [new branch] gh/ydwu4/312/head -> origin/gh/ydwu4/312/head 2025-09-07T07:36:04.2199304Z * [new branch] gh/ydwu4/312/orig -> origin/gh/ydwu4/312/orig 2025-09-07T07:36:04.2201169Z * [new branch] gh/ydwu4/313/base -> origin/gh/ydwu4/313/base 2025-09-07T07:36:04.2202509Z * [new branch] gh/ydwu4/313/head -> origin/gh/ydwu4/313/head 2025-09-07T07:36:04.2204142Z * [new branch] gh/ydwu4/313/orig -> origin/gh/ydwu4/313/orig 2025-09-07T07:36:04.2206184Z * [new branch] gh/ydwu4/314/base -> origin/gh/ydwu4/314/base 2025-09-07T07:36:04.2207489Z * [new branch] gh/ydwu4/314/head -> origin/gh/ydwu4/314/head 2025-09-07T07:36:04.2208702Z * [new branch] gh/ydwu4/314/orig -> origin/gh/ydwu4/314/orig 2025-09-07T07:36:04.2210465Z * [new branch] gh/ydwu4/315/base -> origin/gh/ydwu4/315/base 2025-09-07T07:36:04.2211747Z * [new branch] gh/ydwu4/315/head -> origin/gh/ydwu4/315/head 2025-09-07T07:36:04.2212931Z * [new branch] gh/ydwu4/315/orig -> origin/gh/ydwu4/315/orig 2025-09-07T07:36:04.2214782Z * [new branch] gh/ydwu4/316/base -> origin/gh/ydwu4/316/base 2025-09-07T07:36:04.2215992Z * [new branch] gh/ydwu4/316/head -> origin/gh/ydwu4/316/head 2025-09-07T07:36:04.2217231Z * [new branch] gh/ydwu4/316/orig -> origin/gh/ydwu4/316/orig 2025-09-07T07:36:04.2219217Z * [new branch] gh/ydwu4/317/base -> origin/gh/ydwu4/317/base 2025-09-07T07:36:04.2220214Z * [new branch] gh/ydwu4/317/head -> origin/gh/ydwu4/317/head 2025-09-07T07:36:04.2221462Z * [new branch] gh/ydwu4/317/orig -> origin/gh/ydwu4/317/orig 2025-09-07T07:36:04.2223237Z * [new branch] gh/ydwu4/318/base -> origin/gh/ydwu4/318/base 2025-09-07T07:36:04.2224513Z * [new branch] gh/ydwu4/318/head -> origin/gh/ydwu4/318/head 2025-09-07T07:36:04.2225686Z * [new branch] gh/ydwu4/318/orig -> origin/gh/ydwu4/318/orig 2025-09-07T07:36:04.2227294Z * [new branch] gh/ydwu4/319/base -> origin/gh/ydwu4/319/base 2025-09-07T07:36:04.2228470Z * [new branch] gh/ydwu4/319/head -> origin/gh/ydwu4/319/head 2025-09-07T07:36:04.2229659Z * [new branch] gh/ydwu4/319/orig -> origin/gh/ydwu4/319/orig 2025-09-07T07:36:04.2231663Z * [new branch] gh/ydwu4/320/base -> origin/gh/ydwu4/320/base 2025-09-07T07:36:04.2232804Z * [new branch] gh/ydwu4/320/head -> origin/gh/ydwu4/320/head 2025-09-07T07:36:04.2234025Z * [new branch] gh/ydwu4/320/orig -> origin/gh/ydwu4/320/orig 2025-09-07T07:36:04.2235675Z * [new branch] gh/ydwu4/321/base -> origin/gh/ydwu4/321/base 2025-09-07T07:36:04.2236946Z * [new branch] gh/ydwu4/321/head -> origin/gh/ydwu4/321/head 2025-09-07T07:36:04.2238227Z * [new branch] gh/ydwu4/321/orig -> origin/gh/ydwu4/321/orig 2025-09-07T07:36:04.2239976Z * [new branch] gh/ydwu4/322/base -> origin/gh/ydwu4/322/base 2025-09-07T07:36:04.2241165Z * [new branch] gh/ydwu4/322/head -> origin/gh/ydwu4/322/head 2025-09-07T07:36:04.2242354Z * [new branch] gh/ydwu4/322/orig -> origin/gh/ydwu4/322/orig 2025-09-07T07:36:04.2244100Z * [new branch] gh/ydwu4/323/base -> origin/gh/ydwu4/323/base 2025-09-07T07:36:04.2245270Z * [new branch] gh/ydwu4/323/head -> origin/gh/ydwu4/323/head 2025-09-07T07:36:04.2246436Z * [new branch] gh/ydwu4/323/orig -> origin/gh/ydwu4/323/orig 2025-09-07T07:36:04.2248216Z * [new branch] gh/ydwu4/324/base -> origin/gh/ydwu4/324/base 2025-09-07T07:36:04.2249407Z * [new branch] gh/ydwu4/324/head -> origin/gh/ydwu4/324/head 2025-09-07T07:36:04.2250586Z * [new branch] gh/ydwu4/324/orig -> origin/gh/ydwu4/324/orig 2025-09-07T07:36:04.2252849Z * [new branch] gh/yf225/133/base -> origin/gh/yf225/133/base 2025-09-07T07:36:04.2254069Z * [new branch] gh/yf225/133/head -> origin/gh/yf225/133/head 2025-09-07T07:36:04.2256022Z * [new branch] gh/yf225/171/base -> origin/gh/yf225/171/base 2025-09-07T07:36:04.2257300Z * [new branch] gh/yf225/171/head -> origin/gh/yf225/171/head 2025-09-07T07:36:04.2258534Z * [new branch] gh/yf225/171/orig -> origin/gh/yf225/171/orig 2025-09-07T07:36:04.2260323Z * [new branch] gh/yf225/172/base -> origin/gh/yf225/172/base 2025-09-07T07:36:04.2261450Z * [new branch] gh/yf225/172/head -> origin/gh/yf225/172/head 2025-09-07T07:36:04.2262594Z * [new branch] gh/yf225/172/orig -> origin/gh/yf225/172/orig 2025-09-07T07:36:04.2264290Z * [new branch] gh/yf225/93/base -> origin/gh/yf225/93/base 2025-09-07T07:36:04.2265478Z * [new branch] gh/yf225/93/head -> origin/gh/yf225/93/head 2025-09-07T07:36:04.2268055Z * [new branch] gh/yifuwang/152/base -> origin/gh/yifuwang/152/base 2025-09-07T07:36:04.2269608Z * [new branch] gh/yifuwang/152/head -> origin/gh/yifuwang/152/head 2025-09-07T07:36:04.2271008Z * [new branch] gh/yifuwang/152/orig -> origin/gh/yifuwang/152/orig 2025-09-07T07:36:04.2272587Z * [new branch] gh/yifuwang/195/base -> origin/gh/yifuwang/195/base 2025-09-07T07:36:04.2273862Z * [new branch] gh/yifuwang/195/head -> origin/gh/yifuwang/195/head 2025-09-07T07:36:04.2275072Z * [new branch] gh/yifuwang/195/orig -> origin/gh/yifuwang/195/orig 2025-09-07T07:36:04.2277218Z * [new branch] gh/yiming0416/1/base -> origin/gh/yiming0416/1/base 2025-09-07T07:36:04.2278411Z * [new branch] gh/yiming0416/1/head -> origin/gh/yiming0416/1/head 2025-09-07T07:36:04.2279984Z * [new branch] gh/yiming0416/2/base -> origin/gh/yiming0416/2/base 2025-09-07T07:36:04.2281113Z * [new branch] gh/yiming0416/2/head -> origin/gh/yiming0416/2/head 2025-09-07T07:36:04.2283243Z * [new branch] gh/ysiraichi/79/base -> origin/gh/ysiraichi/79/base 2025-09-07T07:36:04.2284453Z * [new branch] gh/ysiraichi/79/head -> origin/gh/ysiraichi/79/head 2025-09-07T07:36:04.2285889Z * [new branch] gh/ysiraichi/79/orig -> origin/gh/ysiraichi/79/orig 2025-09-07T07:36:04.2287557Z * [new branch] gh/ysiraichi/88/base -> origin/gh/ysiraichi/88/base 2025-09-07T07:36:04.2288737Z * [new branch] gh/ysiraichi/88/head -> origin/gh/ysiraichi/88/head 2025-09-07T07:36:04.2289931Z * [new branch] gh/ysiraichi/88/orig -> origin/gh/ysiraichi/88/orig 2025-09-07T07:36:04.2292222Z * [new branch] gh/zhxchen17/25/base -> origin/gh/zhxchen17/25/base 2025-09-07T07:36:04.2293418Z * [new branch] gh/zhxchen17/25/head -> origin/gh/zhxchen17/25/head 2025-09-07T07:36:04.2294614Z * [new branch] gh/zhxchen17/25/orig -> origin/gh/zhxchen17/25/orig 2025-09-07T07:36:04.2296419Z * [new branch] gh/zhxchen17/31/base -> origin/gh/zhxchen17/31/base 2025-09-07T07:36:04.2297696Z * [new branch] gh/zhxchen17/31/head -> origin/gh/zhxchen17/31/head 2025-09-07T07:36:04.2298957Z * [new branch] gh/zhxchen17/31/orig -> origin/gh/zhxchen17/31/orig 2025-09-07T07:36:04.2300690Z * [new branch] gh/zhxchen17/34/base -> origin/gh/zhxchen17/34/base 2025-09-07T07:36:04.2301972Z * [new branch] gh/zhxchen17/34/head -> origin/gh/zhxchen17/34/head 2025-09-07T07:36:04.2303723Z * [new branch] gh/zhxchen17/35/base -> origin/gh/zhxchen17/35/base 2025-09-07T07:36:04.2304998Z * [new branch] gh/zhxchen17/35/head -> origin/gh/zhxchen17/35/head 2025-09-07T07:36:04.2307087Z * [new branch] gh/zhxchen17/37/base -> origin/gh/zhxchen17/37/base 2025-09-07T07:36:04.2308394Z * [new branch] gh/zhxchen17/37/head -> origin/gh/zhxchen17/37/head 2025-09-07T07:36:04.2309665Z * [new branch] gh/zhxchen17/37/orig -> origin/gh/zhxchen17/37/orig 2025-09-07T07:36:04.2311622Z * [new branch] gh/zhxchen17/38/base -> origin/gh/zhxchen17/38/base 2025-09-07T07:36:04.2312754Z * [new branch] gh/zhxchen17/38/head -> origin/gh/zhxchen17/38/head 2025-09-07T07:36:04.2314011Z * [new branch] gh/zhxchen17/38/orig -> origin/gh/zhxchen17/38/orig 2025-09-07T07:36:04.2316078Z * [new branch] gh/zhxchen17/39/base -> origin/gh/zhxchen17/39/base 2025-09-07T07:36:04.2317435Z * [new branch] gh/zhxchen17/39/head -> origin/gh/zhxchen17/39/head 2025-09-07T07:36:04.2318718Z * [new branch] gh/zhxchen17/39/orig -> origin/gh/zhxchen17/39/orig 2025-09-07T07:36:04.2334778Z * [new branch] gh/zhxchen17/40/base -> origin/gh/zhxchen17/40/base 2025-09-07T07:36:04.2335360Z * [new branch] gh/zhxchen17/40/head -> origin/gh/zhxchen17/40/head 2025-09-07T07:36:04.2336012Z * [new branch] gh/zhxchen17/40/orig -> origin/gh/zhxchen17/40/orig 2025-09-07T07:36:04.2336160Z * [new branch] gh/zhxchen17/41/base -> origin/gh/zhxchen17/41/base 2025-09-07T07:36:04.2336294Z * [new branch] gh/zhxchen17/41/head -> origin/gh/zhxchen17/41/head 2025-09-07T07:36:04.2336431Z * [new branch] gh/zhxchen17/41/orig -> origin/gh/zhxchen17/41/orig 2025-09-07T07:36:04.2336559Z * [new branch] gh/zhxchen17/42/base -> origin/gh/zhxchen17/42/base 2025-09-07T07:36:04.2336688Z * [new branch] gh/zhxchen17/42/head -> origin/gh/zhxchen17/42/head 2025-09-07T07:36:04.2336819Z * [new branch] gh/zhxchen17/42/orig -> origin/gh/zhxchen17/42/orig 2025-09-07T07:36:04.2336952Z * [new branch] gh/zhxchen17/43/base -> origin/gh/zhxchen17/43/base 2025-09-07T07:36:04.2337122Z * [new branch] gh/zhxchen17/43/head -> origin/gh/zhxchen17/43/head 2025-09-07T07:36:04.2337253Z * [new branch] gh/zhxchen17/43/orig -> origin/gh/zhxchen17/43/orig 2025-09-07T07:36:04.2339134Z * [new branch] gh/zhxchen17/44/base -> origin/gh/zhxchen17/44/base 2025-09-07T07:36:04.2340264Z * [new branch] gh/zhxchen17/44/head -> origin/gh/zhxchen17/44/head 2025-09-07T07:36:04.2341454Z * [new branch] gh/zhxchen17/44/orig -> origin/gh/zhxchen17/44/orig 2025-09-07T07:36:04.2343144Z * [new branch] gh/zhxchen17/45/base -> origin/gh/zhxchen17/45/base 2025-09-07T07:36:04.2344639Z * [new branch] gh/zhxchen17/45/head -> origin/gh/zhxchen17/45/head 2025-09-07T07:36:04.2345903Z * [new branch] gh/zhxchen17/45/orig -> origin/gh/zhxchen17/45/orig 2025-09-07T07:36:04.2348042Z * [new branch] gh/zklaus/10/base -> origin/gh/zklaus/10/base 2025-09-07T07:36:04.2349271Z * [new branch] gh/zklaus/10/head -> origin/gh/zklaus/10/head 2025-09-07T07:36:04.2350467Z * [new branch] gh/zklaus/10/orig -> origin/gh/zklaus/10/orig 2025-09-07T07:36:04.2352201Z * [new branch] gh/zklaus/11/base -> origin/gh/zklaus/11/base 2025-09-07T07:36:04.2353370Z * [new branch] gh/zklaus/11/head -> origin/gh/zklaus/11/head 2025-09-07T07:36:04.2354691Z * [new branch] gh/zklaus/11/orig -> origin/gh/zklaus/11/orig 2025-09-07T07:36:04.2356332Z * [new branch] gh/zklaus/12/base -> origin/gh/zklaus/12/base 2025-09-07T07:36:04.2357543Z * [new branch] gh/zklaus/12/head -> origin/gh/zklaus/12/head 2025-09-07T07:36:04.2358704Z * [new branch] gh/zklaus/12/orig -> origin/gh/zklaus/12/orig 2025-09-07T07:36:04.2360367Z * [new branch] gh/zklaus/14/base -> origin/gh/zklaus/14/base 2025-09-07T07:36:04.2361525Z * [new branch] gh/zklaus/14/head -> origin/gh/zklaus/14/head 2025-09-07T07:36:04.2362733Z * [new branch] gh/zklaus/14/orig -> origin/gh/zklaus/14/orig 2025-09-07T07:36:04.2364408Z * [new branch] gh/zklaus/15/base -> origin/gh/zklaus/15/base 2025-09-07T07:36:04.2365666Z * [new branch] gh/zklaus/15/head -> origin/gh/zklaus/15/head 2025-09-07T07:36:04.2366825Z * [new branch] gh/zklaus/15/orig -> origin/gh/zklaus/15/orig 2025-09-07T07:36:04.2368538Z * [new branch] gh/zklaus/16/base -> origin/gh/zklaus/16/base 2025-09-07T07:36:04.2369773Z * [new branch] gh/zklaus/16/head -> origin/gh/zklaus/16/head 2025-09-07T07:36:04.2371001Z * [new branch] gh/zklaus/16/orig -> origin/gh/zklaus/16/orig 2025-09-07T07:36:04.2372746Z * [new branch] gh/zklaus/17/base -> origin/gh/zklaus/17/base 2025-09-07T07:36:04.2373912Z * [new branch] gh/zklaus/17/head -> origin/gh/zklaus/17/head 2025-09-07T07:36:04.2375198Z * [new branch] gh/zklaus/17/orig -> origin/gh/zklaus/17/orig 2025-09-07T07:36:04.2376747Z * [new branch] gh/zklaus/18/base -> origin/gh/zklaus/18/base 2025-09-07T07:36:04.2378060Z * [new branch] gh/zklaus/18/head -> origin/gh/zklaus/18/head 2025-09-07T07:36:04.2379410Z * [new branch] gh/zklaus/18/orig -> origin/gh/zklaus/18/orig 2025-09-07T07:36:04.2381067Z * [new branch] gh/zklaus/19/base -> origin/gh/zklaus/19/base 2025-09-07T07:36:04.2382313Z * [new branch] gh/zklaus/19/head -> origin/gh/zklaus/19/head 2025-09-07T07:36:04.2383469Z * [new branch] gh/zklaus/19/orig -> origin/gh/zklaus/19/orig 2025-09-07T07:36:04.2385140Z * [new branch] gh/zklaus/20/base -> origin/gh/zklaus/20/base 2025-09-07T07:36:04.2386296Z * [new branch] gh/zklaus/20/head -> origin/gh/zklaus/20/head 2025-09-07T07:36:04.2387593Z * [new branch] gh/zklaus/20/orig -> origin/gh/zklaus/20/orig 2025-09-07T07:36:04.2389762Z * [new branch] gh/zklaus/7/base -> origin/gh/zklaus/7/base 2025-09-07T07:36:04.2390950Z * [new branch] gh/zklaus/7/head -> origin/gh/zklaus/7/head 2025-09-07T07:36:04.2392134Z * [new branch] gh/zklaus/7/orig -> origin/gh/zklaus/7/orig 2025-09-07T07:36:04.2393771Z * [new branch] gh/zklaus/9/base -> origin/gh/zklaus/9/base 2025-09-07T07:36:04.2394970Z * [new branch] gh/zklaus/9/head -> origin/gh/zklaus/9/head 2025-09-07T07:36:04.2396139Z * [new branch] gh/zklaus/9/orig -> origin/gh/zklaus/9/orig 2025-09-07T07:36:04.2398328Z * [new branch] gh/zou3519/1175/base -> origin/gh/zou3519/1175/base 2025-09-07T07:36:04.2399505Z * [new branch] gh/zou3519/1175/head -> origin/gh/zou3519/1175/head 2025-09-07T07:36:04.2400701Z * [new branch] gh/zou3519/1175/orig -> origin/gh/zou3519/1175/orig 2025-09-07T07:36:04.2402436Z * [new branch] gh/zou3519/1177/base -> origin/gh/zou3519/1177/base 2025-09-07T07:36:04.2403954Z * [new branch] gh/zou3519/1177/head -> origin/gh/zou3519/1177/head 2025-09-07T07:36:04.2405140Z * [new branch] gh/zou3519/1177/orig -> origin/gh/zou3519/1177/orig 2025-09-07T07:36:04.2406896Z * [new branch] gh/zou3519/1191/base -> origin/gh/zou3519/1191/base 2025-09-07T07:36:04.2408217Z * [new branch] gh/zou3519/1191/head -> origin/gh/zou3519/1191/head 2025-09-07T07:36:04.2409432Z * [new branch] gh/zou3519/1191/orig -> origin/gh/zou3519/1191/orig 2025-09-07T07:36:04.2411272Z * [new branch] gh/zou3519/1192/base -> origin/gh/zou3519/1192/base 2025-09-07T07:36:04.2412486Z * [new branch] gh/zou3519/1192/head -> origin/gh/zou3519/1192/head 2025-09-07T07:36:04.2413687Z * [new branch] gh/zou3519/1192/orig -> origin/gh/zou3519/1192/orig 2025-09-07T07:36:04.2415388Z * [new branch] gh/zou3519/1193/base -> origin/gh/zou3519/1193/base 2025-09-07T07:36:04.2416586Z * [new branch] gh/zou3519/1193/head -> origin/gh/zou3519/1193/head 2025-09-07T07:36:04.2417836Z * [new branch] gh/zou3519/1193/orig -> origin/gh/zou3519/1193/orig 2025-09-07T07:36:04.2419513Z * [new branch] gh/zou3519/1194/base -> origin/gh/zou3519/1194/base 2025-09-07T07:36:04.2420838Z * [new branch] gh/zou3519/1194/head -> origin/gh/zou3519/1194/head 2025-09-07T07:36:04.2422074Z * [new branch] gh/zou3519/1194/orig -> origin/gh/zou3519/1194/orig 2025-09-07T07:36:04.2423844Z * [new branch] gh/zou3519/1195/base -> origin/gh/zou3519/1195/base 2025-09-07T07:36:04.2425559Z * [new branch] gh/zou3519/1195/head -> origin/gh/zou3519/1195/head 2025-09-07T07:36:04.2426413Z * [new branch] gh/zou3519/1195/orig -> origin/gh/zou3519/1195/orig 2025-09-07T07:36:04.2427983Z * [new branch] gh/zou3519/1196/base -> origin/gh/zou3519/1196/base 2025-09-07T07:36:04.2429221Z * [new branch] gh/zou3519/1196/head -> origin/gh/zou3519/1196/head 2025-09-07T07:36:04.2430411Z * [new branch] gh/zou3519/1196/orig -> origin/gh/zou3519/1196/orig 2025-09-07T07:36:04.2431981Z * [new branch] gh/zou3519/1197/base -> origin/gh/zou3519/1197/base 2025-09-07T07:36:04.2433179Z * [new branch] gh/zou3519/1197/head -> origin/gh/zou3519/1197/head 2025-09-07T07:36:04.2434513Z * [new branch] gh/zou3519/1197/orig -> origin/gh/zou3519/1197/orig 2025-09-07T07:36:04.2436743Z * [new branch] gh/zpcore/1/base -> origin/gh/zpcore/1/base 2025-09-07T07:36:04.2437925Z * [new branch] gh/zpcore/1/head -> origin/gh/zpcore/1/head 2025-09-07T07:36:04.2439681Z * [new branch] gh/zpcore/10/base -> origin/gh/zpcore/10/base 2025-09-07T07:36:04.2440730Z * [new branch] gh/zpcore/10/head -> origin/gh/zpcore/10/head 2025-09-07T07:36:04.2441906Z * [new branch] gh/zpcore/10/orig -> origin/gh/zpcore/10/orig 2025-09-07T07:36:04.2444143Z * [new branch] gh/zpcore/11/base -> origin/gh/zpcore/11/base 2025-09-07T07:36:04.2445394Z * [new branch] gh/zpcore/11/head -> origin/gh/zpcore/11/head 2025-09-07T07:36:04.2446598Z * [new branch] gh/zpcore/11/orig -> origin/gh/zpcore/11/orig 2025-09-07T07:36:04.2448532Z * [new branch] gh/zpcore/12/base -> origin/gh/zpcore/12/base 2025-09-07T07:36:04.2449955Z * [new branch] gh/zpcore/12/head -> origin/gh/zpcore/12/head 2025-09-07T07:36:04.2451541Z * [new branch] gh/zpcore/12/orig -> origin/gh/zpcore/12/orig 2025-09-07T07:36:04.2453309Z * [new branch] gh/zpcore/13/base -> origin/gh/zpcore/13/base 2025-09-07T07:36:04.2454650Z * [new branch] gh/zpcore/13/head -> origin/gh/zpcore/13/head 2025-09-07T07:36:04.2455793Z * [new branch] gh/zpcore/13/orig -> origin/gh/zpcore/13/orig 2025-09-07T07:36:04.2457520Z * [new branch] gh/zpcore/14/base -> origin/gh/zpcore/14/base 2025-09-07T07:36:04.2458789Z * [new branch] gh/zpcore/14/head -> origin/gh/zpcore/14/head 2025-09-07T07:36:04.2460599Z * [new branch] gh/zpcore/2/base -> origin/gh/zpcore/2/base 2025-09-07T07:36:04.2461828Z * [new branch] gh/zpcore/2/head -> origin/gh/zpcore/2/head 2025-09-07T07:36:04.2463402Z * [new branch] gh/zpcore/3/base -> origin/gh/zpcore/3/base 2025-09-07T07:36:04.2464572Z * [new branch] gh/zpcore/3/head -> origin/gh/zpcore/3/head 2025-09-07T07:36:04.2466137Z * [new branch] gh/zpcore/4/base -> origin/gh/zpcore/4/base 2025-09-07T07:36:04.2467269Z * [new branch] gh/zpcore/4/head -> origin/gh/zpcore/4/head 2025-09-07T07:36:04.2469285Z * [new branch] gh/zpcore/5/base -> origin/gh/zpcore/5/base 2025-09-07T07:36:04.2470536Z * [new branch] gh/zpcore/5/head -> origin/gh/zpcore/5/head 2025-09-07T07:36:04.2472123Z * [new branch] gh/zpcore/6/base -> origin/gh/zpcore/6/base 2025-09-07T07:36:04.2473329Z * [new branch] gh/zpcore/6/head -> origin/gh/zpcore/6/head 2025-09-07T07:36:04.2474914Z * [new branch] gh/zpcore/7/base -> origin/gh/zpcore/7/base 2025-09-07T07:36:04.2476021Z * [new branch] gh/zpcore/7/head -> origin/gh/zpcore/7/head 2025-09-07T07:36:04.2477710Z * [new branch] gh/zpcore/8/base -> origin/gh/zpcore/8/base 2025-09-07T07:36:04.2478813Z * [new branch] gh/zpcore/8/head -> origin/gh/zpcore/8/head 2025-09-07T07:36:04.2480226Z * [new branch] google-main -> origin/google-main 2025-09-07T07:36:04.2482068Z * [new branch] guangyey/external_stream -> origin/guangyey/external_stream 2025-09-07T07:36:04.2483661Z * [new branch] guangyey/host_alloc -> origin/guangyey/host_alloc 2025-09-07T07:36:04.2484680Z * [new branch] guangyey/reimport -> origin/guangyey/reimport 2025-09-07T07:36:04.2485906Z * [new branch] guangyey/test_2025 -> origin/guangyey/test_2025 2025-09-07T07:36:04.2487913Z * [new branch] guilhermeleobas/cherry-pick-55d87d9dfd9 -> origin/guilhermeleobas/cherry-pick-55d87d9dfd9 2025-09-07T07:36:04.2489760Z * [new branch] haozhe/bf16-dynamic-shape -> origin/haozhe/bf16-dynamic-shape 2025-09-07T07:36:04.2491020Z * [new branch] hc_baseline -> origin/hc_baseline 2025-09-07T07:36:04.2492451Z * [new branch] hf_update -> origin/hf_update 2025-09-07T07:36:04.2493681Z * [new branch] hhh_decomp_mul -> origin/hhh_decomp_mul 2025-09-07T07:36:04.2494944Z * [new branch] hhh_rand -> origin/hhh_rand 2025-09-07T07:36:04.2496682Z * [new branch] hoy/mmsplitk -> origin/hoy/mmsplitk 2025-09-07T07:36:04.2497817Z * [new branch] hoy/triton-PR3973 -> origin/hoy/triton-PR3973 2025-09-07T07:36:04.2499140Z * [new branch] hoy/triton-coalescing-baseline -> origin/hoy/triton-coalescing-baseline 2025-09-07T07:36:04.2500240Z * [new branch] hoy/triton-coalescing-new -> origin/hoy/triton-coalescing-new 2025-09-07T07:36:04.2501356Z * [new branch] hoy/triton-coalescing-vec -> origin/hoy/triton-coalescing-vec 2025-09-07T07:36:04.2502610Z * [new branch] inductordecompfix -> origin/inductordecompfix 2025-09-07T07:36:04.2506135Z * [new branch] inline -> origin/inline 2025-09-07T07:36:04.2507512Z * [new branch] inlining -> origin/inlining 2025-09-07T07:36:04.2508780Z * [new branch] inlining-ezyang -> origin/inlining-ezyang 2025-09-07T07:36:04.2510073Z * [new branch] install-torchao-0.13.0 -> origin/install-torchao-0.13.0 2025-09-07T07:36:04.2511296Z * [new branch] int8_sdpa -> origin/int8_sdpa 2025-09-07T07:36:04.2512584Z * [new branch] invoke-subgraph -> origin/invoke-subgraph 2025-09-07T07:36:04.2513886Z * [new branch] issue#58739 -> origin/issue#58739 2025-09-07T07:36:04.2515915Z * [new branch] jcaip/test-cusparselt-version-0.6.2 -> origin/jcaip/test-cusparselt-version-0.6.2 2025-09-07T07:36:04.2516935Z * [new branch] jcaip/update-cusparselt-0.6.2 -> origin/jcaip/update-cusparselt-0.6.2 2025-09-07T07:36:04.2518658Z * [new branch] jeanschmidt/disable_rocm_build_tests -> origin/jeanschmidt/disable_rocm_build_tests 2025-09-07T07:36:04.2519967Z * [new branch] jithunnair-amd-patch-1 -> origin/jithunnair-amd-patch-1 2025-09-07T07:36:04.2521237Z * [new branch] jithunnair-amd-patch-2 -> origin/jithunnair-amd-patch-2 2025-09-07T07:36:04.2522913Z * [new branch] justinchu/attention-tests -> origin/justinchu/attention-tests 2025-09-07T07:36:04.2524058Z * [new branch] justinchu/native-qdq -> origin/justinchu/native-qdq 2025-09-07T07:36:04.2525447Z * [new branch] justinchu/ort-122 -> origin/justinchu/ort-122 2025-09-07T07:36:04.2527190Z * [new branch] justinchuby/dynamo-true -> origin/justinchuby/dynamo-true 2025-09-07T07:36:04.2529022Z * [new branch] kainan666/xlf_debug -> origin/kainan666/xlf_debug 2025-09-07T07:36:04.2530154Z * [new branch] kainan_test -> origin/kainan_test 2025-09-07T07:36:04.2531419Z * [new branch] learnablebias -> origin/learnablebias 2025-09-07T07:36:04.2533132Z * [new branch] leslie/test_group_gemm_epilogues -> origin/leslie/test_group_gemm_epilogues 2025-09-07T07:36:04.2534904Z * [new branch] lessw2020/fix_cutlass_cache_error -> origin/lessw2020/fix_cutlass_cache_error 2025-09-07T07:36:04.2536540Z * [new branch] liaoxuan/shm_all_reduce -> origin/liaoxuan/shm_all_reduce 2025-09-07T07:36:04.2537737Z * [new branch] liaoxuan/test_fa_disable_softmax -> origin/liaoxuan/test_fa_disable_softmax 2025-09-07T07:36:04.2538867Z * [new branch] liaoxuan/test_int8_sdpa -> origin/liaoxuan/test_int8_sdpa 2025-09-07T07:36:04.2540137Z * [new branch] lintbuilddocker -> origin/lintbuilddocker 2025-09-07T07:36:04.2541316Z * [new branch] llama4-stable -> origin/llama4-stable 2025-09-07T07:36:04.2542632Z * [new branch] logdetfix -> origin/logdetfix 2025-09-07T07:36:04.2544886Z * [new branch] lts/release/1.8 -> origin/lts/release/1.8 2025-09-07T07:36:04.2546648Z * [new branch] lucaskabela/#94773 -> origin/lucaskabela/#94773 2025-09-07T07:36:04.2547816Z * [new branch] lucaskabela/flop_counter -> origin/lucaskabela/flop_counter 2025-09-07T07:36:04.2548939Z * [new branch] lucaskabela/func_under_decomp -> origin/lucaskabela/func_under_decomp 2025-09-07T07:36:04.2550067Z * [new branch] lucaskabela/functional_in_dynamo -> origin/lucaskabela/functional_in_dynamo 2025-09-07T07:36:04.2551251Z * [new branch] lucaskabela/install_params_as_graph_attr -> origin/lucaskabela/install_params_as_graph_attr 2025-09-07T07:36:04.2552319Z * [new branch] lucaskabela/issue_120648 -> origin/lucaskabela/issue_120648 2025-09-07T07:36:04.2553798Z * [new branch] lucaskabela/misc_typing_dynamo -> origin/lucaskabela/misc_typing_dynamo 2025-09-07T07:36:04.2555496Z * [new branch] lucaskabela/parameters_as_graph_attr -> origin/lucaskabela/parameters_as_graph_attr 2025-09-07T07:36:04.2556761Z * [new branch] lucaskabela/remove_aot_dispatcher_metadata -> origin/lucaskabela/remove_aot_dispatcher_metadata 2025-09-07T07:36:04.2557927Z * [new branch] lucaskabela/rnn_decomp -> origin/lucaskabela/rnn_decomp 2025-09-07T07:36:04.2559247Z * [new branch] lucaskabela/typing_backends -> origin/lucaskabela/typing_backends 2025-09-07T07:36:04.2560449Z * [new branch] lucaskabela/typing_symbolic_convert -> origin/lucaskabela/typing_symbolic_convert 2025-09-07T07:36:04.2561732Z * [new branch] lucaskabela/typing_utils_improvements -> origin/lucaskabela/typing_utils_improvements 2025-09-07T07:36:04.2563403Z * [new branch] main -> origin/main 2025-09-07T07:36:04.2564910Z * [new branch] main-enable-b200-distributed-tests -> origin/main-enable-b200-distributed-tests 2025-09-07T07:36:04.2566115Z * [new branch] malfet-patch-1 -> origin/malfet-patch-1 2025-09-07T07:36:04.2567461Z * [new branch] malfet-patch-12 -> origin/malfet-patch-12 2025-09-07T07:36:04.2568784Z * [new branch] malfet-patch-14 -> origin/malfet-patch-14 2025-09-07T07:36:04.2570133Z * [new branch] malfet-patch-6 -> origin/malfet-patch-6 2025-09-07T07:36:04.2571446Z * [new branch] malfet-patch-8 -> origin/malfet-patch-8 2025-09-07T07:36:04.2573419Z * [new branch] malfet/be-move-more-settings-to-checkout-pytorch -> origin/malfet/be-move-more-settings-to-checkout-pytorch 2025-09-07T07:36:04.2574615Z * [new branch] malfet/delete-upsteam-cuda -> origin/malfet/delete-upsteam-cuda 2025-09-07T07:36:04.2575699Z * [new branch] malfet/mps-implement-col2im -> origin/malfet/mps-implement-col2im 2025-09-07T07:36:04.2577386Z * [new branch] manuel/test-ops-common-allow-mps -> origin/manuel/test-ops-common-allow-mps 2025-09-07T07:36:04.2578835Z * [new branch] metascroy-patch-1 -> origin/metascroy-patch-1 2025-09-07T07:36:04.2580577Z * [new branch] mlazos/S429861-debug -> origin/mlazos/S429861-debug 2025-09-07T07:36:04.2581675Z * [new branch] mlazos/aa -> origin/mlazos/aa 2025-09-07T07:36:04.2582823Z * [new branch] mlazos/arg-renames -> origin/mlazos/arg-renames 2025-09-07T07:36:04.2584021Z * [new branch] mlazos/backup-test-branch -> origin/mlazos/backup-test-branch 2025-09-07T07:36:04.2585144Z * [new branch] mlazos/bad-cudagraphs -> origin/mlazos/bad-cudagraphs 2025-09-07T07:36:04.2586289Z * [new branch] mlazos/baseline -> origin/mlazos/baseline 2025-09-07T07:36:04.2587433Z * [new branch] mlazos/baseline-graph-breaks -> origin/mlazos/baseline-graph-breaks 2025-09-07T07:36:04.2588550Z * [new branch] mlazos/beta-tensor -> origin/mlazos/beta-tensor 2025-09-07T07:36:04.2589952Z * [new branch] mlazos/better-msg -> origin/mlazos/better-msg 2025-09-07T07:36:04.2592098Z * [new branch] mlazos/buffers -> origin/mlazos/buffers 2025-09-07T07:36:04.2593276Z * [new branch] mlazos/buffers2 -> origin/mlazos/buffers2 2025-09-07T07:36:04.2594808Z * [new branch] mlazos/buffers3 -> origin/mlazos/buffers3 2025-09-07T07:36:04.2596350Z * [new branch] mlazos/ck2 -> origin/mlazos/ck2 2025-09-07T07:36:04.2597788Z * [new branch] mlazos/combokernels -> origin/mlazos/combokernels 2025-09-07T07:36:04.2599061Z * [new branch] mlazos/ctx-cleanup -> origin/mlazos/ctx-cleanup 2025-09-07T07:36:04.2600178Z * [new branch] mlazos/cuda-cmd-log -> origin/mlazos/cuda-cmd-log 2025-09-07T07:36:04.2601518Z * [new branch] mlazos/cudagraph-tests -> origin/mlazos/cudagraph-tests 2025-09-07T07:36:04.2602755Z * [new branch] mlazos/cudagraphs-measurement -> origin/mlazos/cudagraphs-measurement 2025-09-07T07:36:04.2604206Z * [new branch] mlazos/cutlass-test -> origin/mlazos/cutlass-test 2025-09-07T07:36:04.2605484Z * [new branch] mlazos/cutlass-topo-bug -> origin/mlazos/cutlass-topo-bug 2025-09-07T07:36:04.2606654Z * [new branch] mlazos/data-gather -> origin/mlazos/data-gather 2025-09-07T07:36:04.2607902Z * [new branch] mlazos/data-ptrs2 -> origin/mlazos/data-ptrs2 2025-09-07T07:36:04.2609048Z * [new branch] mlazos/data-ptrs3 -> origin/mlazos/data-ptrs3 2025-09-07T07:36:04.2610284Z * [new branch] mlazos/dataclass-proxy -> origin/mlazos/dataclass-proxy 2025-09-07T07:36:04.2611464Z * [new branch] mlazos/dc-attrs -> origin/mlazos/dc-attrs 2025-09-07T07:36:04.2612765Z * [new branch] mlazos/dc-helion -> origin/mlazos/dc-helion 2025-09-07T07:36:04.2613890Z * [new branch] mlazos/dict-fix -> origin/mlazos/dict-fix 2025-09-07T07:36:04.2615168Z * [new branch] mlazos/disable-closures -> origin/mlazos/disable-closures 2025-09-07T07:36:04.2616402Z * [new branch] mlazos/disable-tf -> origin/mlazos/disable-tf 2025-09-07T07:36:04.2617447Z * [new branch] mlazos/dupe-fix -> origin/mlazos/dupe-fix 2025-09-07T07:36:04.2618918Z * [new branch] mlazos/dyn-batch -> origin/mlazos/dyn-batch 2025-09-07T07:36:04.2620236Z * [new branch] mlazos/evt -> origin/mlazos/evt 2025-09-07T07:36:04.2621452Z * [new branch] mlazos/exp_disable -> origin/mlazos/exp_disable 2025-09-07T07:36:04.2622661Z * [new branch] mlazos/extract-examples -> origin/mlazos/extract-examples 2025-09-07T07:36:04.2623827Z * [new branch] mlazos/foreach-op -> origin/mlazos/foreach-op 2025-09-07T07:36:04.2625017Z * [new branch] mlazos/fp8 -> origin/mlazos/fp8 2025-09-07T07:36:04.2626404Z * [new branch] mlazos/fp8-bias -> origin/mlazos/fp8-bias 2025-09-07T07:36:04.2627632Z * [new branch] mlazos/fp8-bias-fusion -> origin/mlazos/fp8-bias-fusion 2025-09-07T07:36:04.2628680Z * [new branch] mlazos/fp8-fixes -> origin/mlazos/fp8-fixes 2025-09-07T07:36:04.2629905Z * [new branch] mlazos/freezing -> origin/mlazos/freezing 2025-09-07T07:36:04.2631143Z * [new branch] mlazos/h-comp -> origin/mlazos/h-comp 2025-09-07T07:36:04.2632427Z * [new branch] mlazos/h-comp2 -> origin/mlazos/h-comp2 2025-09-07T07:36:04.2633661Z * [new branch] mlazos/hash-hop -> origin/mlazos/hash-hop 2025-09-07T07:36:04.2634897Z * [new branch] mlazos/hc -> origin/mlazos/hc 2025-09-07T07:36:04.2636183Z * [new branch] mlazos/hc-cycles -> origin/mlazos/hc-cycles 2025-09-07T07:36:04.2637800Z * [new branch] mlazos/hc-fixes -> origin/mlazos/hc-fixes 2025-09-07T07:36:04.2639091Z * [new branch] mlazos/hc-fixes3 -> origin/mlazos/hc-fixes3 2025-09-07T07:36:04.2640302Z * [new branch] mlazos/hc-fixes4 -> origin/mlazos/hc-fixes4 2025-09-07T07:36:04.2641567Z * [new branch] mlazos/hc-hf -> origin/mlazos/hc-hf 2025-09-07T07:36:04.2642746Z * [new branch] mlazos/hc-mut -> origin/mlazos/hc-mut 2025-09-07T07:36:04.2644069Z * [new branch] mlazos/hc10 -> origin/mlazos/hc10 2025-09-07T07:36:04.2645284Z * [new branch] mlazos/hc11 -> origin/mlazos/hc11 2025-09-07T07:36:04.2646500Z * [new branch] mlazos/hc12 -> origin/mlazos/hc12 2025-09-07T07:36:04.2647679Z * [new branch] mlazos/hc13 -> origin/mlazos/hc13 2025-09-07T07:36:04.2648963Z * [new branch] mlazos/hc14 -> origin/mlazos/hc14 2025-09-07T07:36:04.2650178Z * [new branch] mlazos/hc15 -> origin/mlazos/hc15 2025-09-07T07:36:04.2651440Z * [new branch] mlazos/hc2 -> origin/mlazos/hc2 2025-09-07T07:36:04.2652631Z * [new branch] mlazos/hc4 -> origin/mlazos/hc4 2025-09-07T07:36:04.2653866Z * [new branch] mlazos/hc5 -> origin/mlazos/hc5 2025-09-07T07:36:04.2655078Z * [new branch] mlazos/hc6 -> origin/mlazos/hc6 2025-09-07T07:36:04.2656300Z * [new branch] mlazos/hc7 -> origin/mlazos/hc7 2025-09-07T07:36:04.2657484Z * [new branch] mlazos/hc8 -> origin/mlazos/hc8 2025-09-07T07:36:04.2658889Z * [new branch] mlazos/hc9 -> origin/mlazos/hc9 2025-09-07T07:36:04.2660134Z * [new branch] mlazos/hc_baseline2 -> origin/mlazos/hc_baseline2 2025-09-07T07:36:04.2661342Z * [new branch] mlazos/init-per-param -> origin/mlazos/init-per-param 2025-09-07T07:36:04.2662561Z * [new branch] mlazos/init_per_param -> origin/mlazos/init_per_param 2025-09-07T07:36:04.2663719Z * [new branch] mlazos/less-guards -> origin/mlazos/less-guards 2025-09-07T07:36:04.2664956Z * [new branch] mlazos/lr-composibility -> origin/mlazos/lr-composibility 2025-09-07T07:36:04.2666116Z * [new branch] mlazos/main -> origin/mlazos/main 2025-09-07T07:36:04.2667382Z * [new branch] mlazos/main-test-enablement -> origin/mlazos/main-test-enablement 2025-09-07T07:36:04.2668542Z * [new branch] mlazos/main2 -> origin/mlazos/main2 2025-09-07T07:36:04.2669799Z * [new branch] mlazos/mark-static-update -> origin/mlazos/mark-static-update 2025-09-07T07:36:04.2671031Z * [new branch] mlazos/mcg -> origin/mlazos/mcg 2025-09-07T07:36:04.2672663Z * [new branch] mlazos/mcg2 -> origin/mlazos/mcg2 2025-09-07T07:36:04.2673947Z * [new branch] mlazos/meta-guards -> origin/mlazos/meta-guards 2025-09-07T07:36:04.2675542Z * [new branch] mlazos/mlazos/ck2 -> origin/mlazos/mlazos/ck2 2025-09-07T07:36:04.2676846Z * [new branch] mlazos/mlazos/foreach-map-adam -> origin/mlazos/mlazos/foreach-map-adam 2025-09-07T07:36:04.2678503Z * [new branch] mlazos/mlazos/tf-mode-backup -> origin/mlazos/mlazos/tf-mode-backup 2025-09-07T07:36:04.2679693Z * [new branch] mlazos/mod-fix -> origin/mlazos/mod-fix 2025-09-07T07:36:04.2680938Z * [new branch] mlazos/mode-fix -> origin/mlazos/mode-fix 2025-09-07T07:36:04.2682189Z * [new branch] mlazos/more-tests -> origin/mlazos/more-tests 2025-09-07T07:36:04.2683404Z * [new branch] mlazos/no-cpp -> origin/mlazos/no-cpp 2025-09-07T07:36:04.2684808Z * [new branch] mlazos/no-init-group-handling -> origin/mlazos/no-init-group-handling 2025-09-07T07:36:04.2685948Z * [new branch] mlazos/offsets -> origin/mlazos/offsets 2025-09-07T07:36:04.2687123Z * [new branch] mlazos/opt-bench-exp2 -> origin/mlazos/opt-bench-exp2 2025-09-07T07:36:04.2688441Z * [new branch] mlazos/opt-incr -> origin/mlazos/opt-incr 2025-09-07T07:36:04.2689653Z * [new branch] mlazos/proxy-ctors -> origin/mlazos/proxy-ctors 2025-09-07T07:36:04.2691011Z * [new branch] mlazos/quant-fix -> origin/mlazos/quant-fix 2025-09-07T07:36:04.2692265Z * [new branch] mlazos/resnet-fix -> origin/mlazos/resnet-fix 2025-09-07T07:36:04.2693479Z * [new branch] mlazos/revert-inline -> origin/mlazos/revert-inline 2025-09-07T07:36:04.2694705Z * [new branch] mlazos/rm-buf-names -> origin/mlazos/rm-buf-names 2025-09-07T07:36:04.2695825Z * [new branch] mlazos/rm-code -> origin/mlazos/rm-code 2025-09-07T07:36:04.2697055Z * [new branch] mlazos/rm-spam -> origin/mlazos/rm-spam 2025-09-07T07:36:04.2698417Z * [new branch] mlazos/rtp -> origin/mlazos/rtp 2025-09-07T07:36:04.2699722Z * [new branch] mlazos/static-idx-dbg -> origin/mlazos/static-idx-dbg 2025-09-07T07:36:04.2700941Z * [new branch] mlazos/static-inputs-log -> origin/mlazos/static-inputs-log 2025-09-07T07:36:04.2702181Z * [new branch] mlazos/sub-param-fix -> origin/mlazos/sub-param-fix 2025-09-07T07:36:04.2703561Z * [new branch] mlazos/td-fix2 -> origin/mlazos/td-fix2 2025-09-07T07:36:04.2704879Z * [new branch] mlazos/tensor-hasattr2 -> origin/mlazos/tensor-hasattr2 2025-09-07T07:36:04.2706070Z * [new branch] mlazos/test -> origin/mlazos/test 2025-09-07T07:36:04.2707256Z * [new branch] mlazos/tf-mode -> origin/mlazos/tf-mode 2025-09-07T07:36:04.2708511Z * [new branch] mlazos/tf-mode-backup2 -> origin/mlazos/tf-mode-backup2 2025-09-07T07:36:04.2709728Z * [new branch] mlazos/tf-mode-reland -> origin/mlazos/tf-mode-reland 2025-09-07T07:36:04.2711097Z * [new branch] mlazos/tf-mode-reland2 -> origin/mlazos/tf-mode-reland2 2025-09-07T07:36:04.2712394Z * [new branch] mlazos/tf-mode-reland3 -> origin/mlazos/tf-mode-reland3 2025-09-07T07:36:04.2713492Z * [new branch] mlazos/topo-fix -> origin/mlazos/topo-fix 2025-09-07T07:36:04.2714706Z * [new branch] mlazos/triton-no-epi -> origin/mlazos/triton-no-epi 2025-09-07T07:36:04.2715917Z * [new branch] mlazos/tune-proto -> origin/mlazos/tune-proto 2025-09-07T07:36:04.2717109Z * [new branch] mlazos/tuple-fixes -> origin/mlazos/tuple-fixes 2025-09-07T07:36:04.2718323Z * [new branch] mlazos/tuple-fixes2 -> origin/mlazos/tuple-fixes2 2025-09-07T07:36:04.2719609Z * [new branch] mlazos/tuple-handling -> origin/mlazos/tuple-handling 2025-09-07T07:36:04.2720868Z * [new branch] mlazos/user-streams -> origin/mlazos/user-streams 2025-09-07T07:36:04.2722615Z * [new branch] mlazos/vary-beta -> origin/mlazos/vary-beta 2025-09-07T07:36:04.2723863Z * [new branch] mlazos/vary-beta2 -> origin/mlazos/vary-beta2 2025-09-07T07:36:04.2725157Z * [new branch] mlazos/weird-perf1 -> origin/mlazos/weird-perf1 2025-09-07T07:36:04.2726508Z * [new branch] mm_out_dtype_compile -> origin/mm_out_dtype_compile 2025-09-07T07:36:04.2727763Z * [new branch] modify-setupvllm -> origin/modify-setupvllm 2025-09-07T07:36:04.2729012Z * [new branch] module-shim -> origin/module-shim 2025-09-07T07:36:04.2730362Z * [new branch] move-theme-out-docker -> origin/move-theme-out-docker 2025-09-07T07:36:04.2732093Z * [new branch] msaroufim/be1 -> origin/msaroufim/be1 2025-09-07T07:36:04.2733312Z * [new branch] msaroufim/cn_path -> origin/msaroufim/cn_path 2025-09-07T07:36:04.2734556Z * [new branch] msaroufim/dtensorfusedadam -> origin/msaroufim/dtensorfusedadam 2025-09-07T07:36:04.2735742Z * [new branch] msaroufim/reduce -> origin/msaroufim/reduce 2025-09-07T07:36:04.2737522Z * [new branch] mtia/basic-cmake -> origin/mtia/basic-cmake 2025-09-07T07:36:04.2738892Z * [new branch] muon_dev -> origin/muon_dev 2025-09-07T07:36:04.2740222Z * [new branch] muon_dev_1 -> origin/muon_dev_1 2025-09-07T07:36:04.2741494Z * [new branch] nativert_num_outputs -> origin/nativert_num_outputs 2025-09-07T07:36:04.2742849Z * [new branch] nativert_numoutputs -> origin/nativert_numoutputs 2025-09-07T07:36:04.2744123Z * [new branch] new-modifiy-setupvllm -> origin/new-modifiy-setupvllm 2025-09-07T07:36:04.2745357Z * [new branch] new-setupvllm -> origin/new-setupvllm 2025-09-07T07:36:04.2746637Z * [new branch] new_zeros_dtype -> origin/new_zeros_dtype 2025-09-07T07:36:04.2747982Z * [new branch] newtest-base -> origin/newtest-base 2025-09-07T07:36:04.2749684Z * [new branch] ngimel/cat_perf1 -> origin/ngimel/cat_perf1 2025-09-07T07:36:04.2750819Z * [new branch] ngimel/einsum_fix -> origin/ngimel/einsum_fix 2025-09-07T07:36:04.2752403Z * [new branch] ngimel/error_index_list -> origin/ngimel/error_index_list 2025-09-07T07:36:04.2753547Z * [new branch] ngimel/fabric_check -> origin/ngimel/fabric_check 2025-09-07T07:36:04.2754694Z * [new branch] ngimel/fabric_fix -> origin/ngimel/fabric_fix 2025-09-07T07:36:04.2755989Z * [new branch] ngimel/fix_driver_init_error -> origin/ngimel/fix_driver_init_error 2025-09-07T07:36:04.2757483Z * [new branch] ngimel/fix_nccl_segment_seg -> origin/ngimel/fix_nccl_segment_seg 2025-09-07T07:36:04.2758932Z * [new branch] ngimel/gg_new -> origin/ngimel/gg_new 2025-09-07T07:36:04.2760331Z * [new branch] ngimel/modeguard -> origin/ngimel/modeguard 2025-09-07T07:36:04.2762287Z * [new branch] ngimel/multicast_fix -> origin/ngimel/multicast_fix 2025-09-07T07:36:04.2763630Z * [new branch] ngimel/rocm_handle_type -> origin/ngimel/rocm_handle_type 2025-09-07T07:36:04.2764921Z * [new branch] ngimel/symm_handle_fabric -> origin/ngimel/symm_handle_fabric 2025-09-07T07:36:04.2766128Z * [new branch] ngimel/unbind_multimem -> origin/ngimel/unbind_multimem 2025-09-07T07:36:04.2767398Z * [new branch] nightly -> origin/nightly 2025-09-07T07:36:04.2768767Z * [new branch] nmacchioni-patch-10 -> origin/nmacchioni-patch-10 2025-09-07T07:36:04.2770069Z * [new branch] nmacchioni-patch-7 -> origin/nmacchioni-patch-7 2025-09-07T07:36:04.2771452Z * [new branch] nmacchioni-patch-8 -> origin/nmacchioni-patch-8 2025-09-07T07:36:04.2772843Z * [new branch] nmacchioni-patch-9 -> origin/nmacchioni-patch-9 2025-09-07T07:36:04.2774634Z * [new branch] nullplay/fuse_matmul -> origin/nullplay/fuse_matmul 2025-09-07T07:36:04.2775875Z * [new branch] nullplay_fuse_matmul -> origin/nullplay_fuse_matmul 2025-09-07T07:36:04.2777134Z * [new branch] one-off -> origin/one-off 2025-09-07T07:36:04.2779532Z * [new branch] orig/release/1.10 -> origin/orig/release/1.10 2025-09-07T07:36:04.2780785Z * [new branch] orig/release/1.11 -> origin/orig/release/1.11 2025-09-07T07:36:04.2782034Z * [new branch] orig/release/1.12 -> origin/orig/release/1.12 2025-09-07T07:36:04.2783505Z * [new branch] orig/release/1.13 -> origin/orig/release/1.13 2025-09-07T07:36:04.2784781Z * [new branch] orig/release/1.6 -> origin/orig/release/1.6 2025-09-07T07:36:04.2786120Z * [new branch] orig/release/1.7 -> origin/orig/release/1.7 2025-09-07T07:36:04.2787370Z * [new branch] orig/release/1.8 -> origin/orig/release/1.8 2025-09-07T07:36:04.2788639Z * [new branch] orig/release/1.9 -> origin/orig/release/1.9 2025-09-07T07:36:04.2789901Z * [new branch] orig/release/2.0 -> origin/orig/release/2.0 2025-09-07T07:36:04.2791232Z * [new branch] orig/release/2.1 -> origin/orig/release/2.1 2025-09-07T07:36:04.2792503Z * [new branch] orig/release/2.2 -> origin/orig/release/2.2 2025-09-07T07:36:04.2793686Z * [new branch] orig/release/2.3 -> origin/orig/release/2.3 2025-09-07T07:36:04.2794878Z * [new branch] orig/release/2.4 -> origin/orig/release/2.4 2025-09-07T07:36:04.2796059Z * [new branch] orig/release/2.5 -> origin/orig/release/2.5 2025-09-07T07:36:04.2797279Z * [new branch] orig/release/2.6 -> origin/orig/release/2.6 2025-09-07T07:36:04.2798791Z * [new branch] orig/release/2.7 -> origin/orig/release/2.7 2025-09-07T07:36:04.2800431Z * [new branch] orig/release/2.8 -> origin/orig/release/2.8 2025-09-07T07:36:04.2802269Z * [new branch] oulgen/fx_graph -> origin/oulgen/fx_graph 2025-09-07T07:36:04.2803772Z * [new branch] padded-tensor -> origin/padded-tensor 2025-09-07T07:36:04.2805103Z * [new branch] pca2 -> origin/pca2 2025-09-07T07:36:04.2806504Z * [new branch] pianpwk-patch-1 -> origin/pianpwk-patch-1 2025-09-07T07:36:04.2808297Z * [new branch] pianpwk/backed_size_oblivious_export -> origin/pianpwk/backed_size_oblivious_export 2025-09-07T07:36:04.2809560Z * [new branch] pianpwk/invalidate_fake_memo -> origin/pianpwk/invalidate_fake_memo 2025-09-07T07:36:04.2810584Z * [new branch] pianpwk/max_1_strides -> origin/pianpwk/max_1_strides 2025-09-07T07:36:04.2811714Z * [new branch] pianpwk/maybe_guard_rel -> origin/pianpwk/maybe_guard_rel 2025-09-07T07:36:04.2812803Z * [new branch] pianpwk/nonzero_memo -> origin/pianpwk/nonzero_memo 2025-09-07T07:36:04.2814260Z * [new branch] pianpwk/oblivious_reshape_view_better -> origin/pianpwk/oblivious_reshape_view_better 2025-09-07T07:36:04.2815758Z * [new branch] pianpwk/oblivious_slice_forward -> origin/pianpwk/oblivious_slice_forward 2025-09-07T07:36:04.2816998Z * [new branch] pianpwk/oblivious_where -> origin/pianpwk/oblivious_where 2025-09-07T07:36:04.2818281Z * [new branch] pianpwk/param_static_pgo -> origin/pianpwk/param_static_pgo 2025-09-07T07:36:04.2819478Z * [new branch] pianpwk/pre_forward_hook -> origin/pianpwk/pre_forward_hook 2025-09-07T07:36:04.2820840Z * [new branch] pianpwk/remove_guard_fail_break -> origin/pianpwk/remove_guard_fail_break 2025-09-07T07:36:04.2822014Z * [new branch] pianpwk/slice_fresh_symbols -> origin/pianpwk/slice_fresh_symbols 2025-09-07T07:36:04.2823178Z * [new branch] pianpwk/sym_tokens_draft -> origin/pianpwk/sym_tokens_draft 2025-09-07T07:36:04.2824638Z * [new branch] pianpwk/test_pointwise_guard_or_false -> origin/pianpwk/test_pointwise_guard_or_false 2025-09-07T07:36:04.2825815Z * [new branch] pianpwk/test_slice_fake_impl -> origin/pianpwk/test_slice_fake_impl 2025-09-07T07:36:04.2827076Z * [new branch] pianpwk/totally_draft_sym_wrap -> origin/pianpwk/totally_draft_sym_wrap 2025-09-07T07:36:04.2828271Z * [new branch] pianpwk/unbacked_channels_last -> origin/pianpwk/unbacked_channels_last 2025-09-07T07:36:04.2829556Z * [new branch] pianpwk/unbacked_safe_conv1d -> origin/pianpwk/unbacked_safe_conv1d 2025-09-07T07:36:04.2830774Z * [new branch] pianpwk/unbacked_sdpa_flash -> origin/pianpwk/unbacked_sdpa_flash 2025-09-07T07:36:04.2832079Z * [new branch] pianpwk/unbacked_should_swap -> origin/pianpwk/unbacked_should_swap 2025-09-07T07:36:04.2833256Z * [new branch] pianpwk/unbacked_should_swap_2 -> origin/pianpwk/unbacked_should_swap_2 2025-09-07T07:36:04.2834441Z * [new branch] pianpwk/unbacked_slice_binding -> origin/pianpwk/unbacked_slice_binding 2025-09-07T07:36:04.2835642Z * [new branch] pianpwk/unbacked_slice_forward -> origin/pianpwk/unbacked_slice_forward 2025-09-07T07:36:04.2836876Z * [new branch] pianpwk/user_symints -> origin/pianpwk/user_symints 2025-09-07T07:36:04.2839040Z * [new branch] pianpwk/wan21_reshape -> origin/pianpwk/wan21_reshape 2025-09-07T07:36:04.2840519Z * [new branch] pianpwk/whitelist_optimizer -> origin/pianpwk/whitelist_optimizer 2025-09-07T07:36:04.2841153Z * [new branch] pin-torchao -> origin/pin-torchao 2025-09-07T07:36:04.2842572Z * [new branch] piz/fall_back_missing_0716 -> origin/piz/fall_back_missing_0716 2025-09-07T07:36:04.2843680Z * [new branch] piz/improve_scatter_0808 -> origin/piz/improve_scatter_0808 2025-09-07T07:36:04.2844966Z * [new branch] pool-separate -> origin/pool-separate 2025-09-07T07:36:04.2846226Z * [new branch] pr-156087 -> origin/pr-156087 2025-09-07T07:36:04.2848061Z * [new branch] pr/131860 -> origin/pr/131860 2025-09-07T07:36:04.2849351Z * [new branch] predispatch_to -> origin/predispatch_to 2025-09-07T07:36:04.2850698Z * [new branch] pt-opt-cuda3 -> origin/pt-opt-cuda3 2025-09-07T07:36:04.2851974Z * [new branch] pyobjectslot -> origin/pyobjectslot 2025-09-07T07:36:04.2854032Z * [new branch] python_compiled_autograd -> origin/python_compiled_autograd 2025-09-07T07:36:04.2856226Z * [new branch] qchip/export-D54134695 -> origin/qchip/export-D54134695 2025-09-07T07:36:04.2857470Z * [new branch] quint-bits -> origin/quint-bits 2025-09-07T07:36:04.2859983Z * [new branch] release/1.10 -> origin/release/1.10 2025-09-07T07:36:04.2861252Z * [new branch] release/1.11 -> origin/release/1.11 2025-09-07T07:36:04.2862498Z * [new branch] release/1.12 -> origin/release/1.12 2025-09-07T07:36:04.2863719Z * [new branch] release/1.13 -> origin/release/1.13 2025-09-07T07:36:04.2864909Z * [new branch] release/1.4 -> origin/release/1.4 2025-09-07T07:36:04.2865942Z * [new branch] release/1.4.1 -> origin/release/1.4.1 2025-09-07T07:36:04.2867160Z * [new branch] release/1.5 -> origin/release/1.5 2025-09-07T07:36:04.2868434Z * [new branch] release/1.6 -> origin/release/1.6 2025-09-07T07:36:04.2869678Z * [new branch] release/1.7 -> origin/release/1.7 2025-09-07T07:36:04.2871057Z * [new branch] release/1.8 -> origin/release/1.8 2025-09-07T07:36:04.2872183Z * [new branch] release/1.9 -> origin/release/1.9 2025-09-07T07:36:04.2873420Z * [new branch] release/2.0 -> origin/release/2.0 2025-09-07T07:36:04.2874714Z * [new branch] release/2.1 -> origin/release/2.1 2025-09-07T07:36:04.2875975Z * [new branch] release/2.2 -> origin/release/2.2 2025-09-07T07:36:04.2877526Z * [new branch] release/2.3 -> origin/release/2.3 2025-09-07T07:36:04.2879209Z * [new branch] release/2.4 -> origin/release/2.4 2025-09-07T07:36:04.2880848Z * [new branch] release/2.5 -> origin/release/2.5 2025-09-07T07:36:04.2882238Z * [new branch] release/2.6 -> origin/release/2.6 2025-09-07T07:36:04.2883528Z * [new branch] release/2.7 -> origin/release/2.7 2025-09-07T07:36:04.2884796Z * [new branch] release/2.8 -> origin/release/2.8 2025-09-07T07:36:04.2886106Z * [new branch] release_notes -> origin/release_notes 2025-09-07T07:36:04.2887416Z * [new branch] remove-actionable-label -> origin/remove-actionable-label 2025-09-07T07:36:04.2888649Z * [new branch] remove-ao -> origin/remove-ao 2025-09-07T07:36:04.2890160Z * [new branch] removedeprecatedvllmtest -> origin/removedeprecatedvllmtest 2025-09-07T07:36:04.2891529Z * [new branch] replace-pytorch-labs-20250812-195836 -> origin/replace-pytorch-labs-20250812-195836 2025-09-07T07:36:04.2892655Z * [new branch] replace-pytorch-labs-20250812-200248 -> origin/replace-pytorch-labs-20250812-200248 2025-09-07T07:36:04.2893904Z * [new branch] replace-pytorch-labs-20250812-200324 -> origin/replace-pytorch-labs-20250812-200324 2025-09-07T07:36:04.2895214Z * [new branch] replace-pytorch-labs-20250812-204020 -> origin/replace-pytorch-labs-20250812-204020 2025-09-07T07:36:04.2896438Z * [new branch] replace-pytorch-labs-20250812-204125 -> origin/replace-pytorch-labs-20250812-204125 2025-09-07T07:36:04.2897853Z * [new branch] replace-pytorch-labs-20250812-205624 -> origin/replace-pytorch-labs-20250812-205624 2025-09-07T07:36:04.2900593Z * [new branch] revert-131069-gh/krzysztofjordan/1/head -> origin/revert-131069-gh/krzysztofjordan/1/head 2025-09-07T07:36:04.2903331Z * [new branch] revert-131469-gh/andrewor14/51/head -> origin/revert-131469-gh/andrewor14/51/head 2025-09-07T07:36:04.2906190Z * [new branch] revert-156870-gh/skarjala/3/head -> origin/revert-156870-gh/skarjala/3/head 2025-09-07T07:36:04.2907562Z * [new branch] revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ -> origin/revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ 2025-09-07T07:36:04.2908655Z * [new branch] rocm-monitoring -> origin/rocm-monitoring 2025-09-07T07:36:04.2910458Z * [new branch] ruisi/relax_memory -> origin/ruisi/relax_memory 2025-09-07T07:36:04.2911776Z * [new branch] run-torchbench-smoke-test-h100 -> origin/run-torchbench-smoke-test-h100 2025-09-07T07:36:04.2913677Z * [new branch] ryanguo99/cleanup-dynamo-expected-failures -> origin/ryanguo99/cleanup-dynamo-expected-failures 2025-09-07T07:36:04.2914741Z * [new branch] ryanguo99/fix-closure-var -> origin/ryanguo99/fix-closure-var 2025-09-07T07:36:04.2916474Z * [new branch] rzou/faketensor_bench -> origin/rzou/faketensor_bench 2025-09-07T07:36:04.2917597Z * [new branch] rzou/njt -> origin/rzou/njt 2025-09-07T07:36:04.2918783Z * [new branch] rzou/pca -> origin/rzou/pca 2025-09-07T07:36:04.2919928Z * [new branch] rzou/realprop -> origin/rzou/realprop 2025-09-07T07:36:04.2921121Z * [new branch] rzou/setup_context -> origin/rzou/setup_context 2025-09-07T07:36:04.2923069Z * [new branch] sanchitintel/refactor_aten_int8_woq_gemm -> origin/sanchitintel/refactor_aten_int8_woq_gemm 2025-09-07T07:36:04.2924329Z * [new branch] sanchitintel/weird_thing_with_test_cpu_select_algorithm -> origin/sanchitintel/weird_thing_with_test_cpu_select_algorithm 2025-09-07T07:36:04.2925558Z * [new branch] sapling-pr-archive-SS-JIA -> origin/sapling-pr-archive-SS-JIA 2025-09-07T07:36:04.2926775Z * [new branch] save -> origin/save 2025-09-07T07:36:04.2928600Z * [new branch] sdym/2.5.1 -> origin/sdym/2.5.1 2025-09-07T07:36:04.2929926Z * [new branch] seemethere-patch-1 -> origin/seemethere-patch-1 2025-09-07T07:36:04.2931251Z * [new branch] setupvllm -> origin/setupvllm 2025-09-07T07:36:04.2932656Z * [new branch] share_and_pin_fork -> origin/share_and_pin_fork 2025-09-07T07:36:04.2934489Z * [new branch] shengf/fx-xform-perf -> origin/shengf/fx-xform-perf 2025-09-07T07:36:04.2935716Z * [new branch] shikaili_fp8_allgather -> origin/shikaili_fp8_allgather 2025-09-07T07:36:04.2937065Z * [new branch] shoumikhin-patch-1 -> origin/shoumikhin-patch-1 2025-09-07T07:36:04.2938588Z * [new branch] shoumikhin-patch-12 -> origin/shoumikhin-patch-12 2025-09-07T07:36:04.2940131Z * [new branch] simplify-fq-per-channel -> origin/simplify-fq-per-channel 2025-09-07T07:36:04.2941470Z * [new branch] solve-accuracy-fix -> origin/solve-accuracy-fix 2025-09-07T07:36:04.2943130Z * [new branch] soulitzer/stash-tls-ac -> origin/soulitzer/stash-tls-ac 2025-09-07T07:36:04.2944895Z * [new branch] sqzhang/flight4 -> origin/sqzhang/flight4 2025-09-07T07:36:04.2946128Z * [new branch] sqzhang/flight4plus -> origin/sqzhang/flight4plus 2025-09-07T07:36:04.2947866Z * [new branch] sraikund/record_funct_test -> origin/sraikund/record_funct_test 2025-09-07T07:36:04.2949575Z * [new branch] sraikund16/test -> origin/sraikund16/test 2025-09-07T07:36:04.2951431Z * [new branch] stablize-compilation-time -> origin/stablize-compilation-time 2025-09-07T07:36:04.2952693Z * [new branch] standalone-templates -> origin/standalone-templates 2025-09-07T07:36:04.2954104Z * [new branch] standalone_package_weights -> origin/standalone_package_weights 2025-09-07T07:36:04.2955240Z * [new branch] starterTaskUpdate -> origin/starterTaskUpdate 2025-09-07T07:36:04.2956475Z * [new branch] subgraph_fuse -> origin/subgraph_fuse 2025-09-07T07:36:04.2957845Z * [new branch] support-uv-in-collect_env -> origin/support-uv-in-collect_env 2025-09-07T07:36:04.2959070Z * [new branch] sve-poc -> origin/sve-poc 2025-09-07T07:36:04.2960332Z * [new branch] svekars-patch-1 -> origin/svekars-patch-1 2025-09-07T07:36:04.2961773Z * [new branch] switch-bn -> origin/switch-bn 2025-09-07T07:36:04.2963065Z * [new branch] sympy-bottleneck-repro -> origin/sympy-bottleneck-repro 2025-09-07T07:36:04.2964870Z * [new branch] tenpercent/ck_rocm_ci_v3 -> origin/tenpercent/ck_rocm_ci_v3 2025-09-07T07:36:04.2966233Z * [new branch] tensordict_integration -> origin/tensordict_integration 2025-09-07T07:36:04.2967456Z * [new branch] test-7054 -> origin/test-7054 2025-09-07T07:36:04.2968935Z * [new branch] test-move-conda-builds -> origin/test-move-conda-builds 2025-09-07T07:36:04.2970348Z * [new branch] test-myst-markdown-docstring -> origin/test-myst-markdown-docstring 2025-09-07T07:36:04.2971551Z * [new branch] test-old -> origin/test-old 2025-09-07T07:36:04.2972866Z * [new branch] test-vec-migration-internally -> origin/test-vec-migration-internally 2025-09-07T07:36:04.2974513Z * [new branch] test/bmm_heur -> origin/test/bmm_heur 2025-09-07T07:36:04.2975685Z * [new branch] test/inductor -> origin/test/inductor 2025-09-07T07:36:04.2977452Z * [new branch] tianren/flex_paged_attn_fix -> origin/tianren/flex_paged_attn_fix 2025-09-07T07:36:04.2978712Z * [new branch] tianren/flex_paged_attn_fix_temp -> origin/tianren/flex_paged_attn_fix_temp 2025-09-07T07:36:04.2979783Z * [new branch] tianren/test -> origin/tianren/test 2025-09-07T07:36:04.2981080Z * [new branch] tidy_performance_cyy -> origin/tidy_performance_cyy 2025-09-07T07:36:04.2982371Z * [new branch] torchtitan_ep -> origin/torchtitan_ep 2025-09-07T07:36:04.2983844Z * [new branch] trace_fsdp_torchtune_lora -> origin/trace_fsdp_torchtune_lora 2025-09-07T07:36:04.2985044Z * [new branch] traceable_fsdp_unit_tests -> origin/traceable_fsdp_unit_tests 2025-09-07T07:36:04.2986362Z * [new branch] tree_loop_vec_base -> origin/tree_loop_vec_base 2025-09-07T07:36:04.2987667Z * [new branch] tree_vec_base -> origin/tree_vec_base 2025-09-07T07:36:04.2989002Z * [new branch] triton-update -> origin/triton-update 2025-09-07T07:36:04.2990229Z * [new branch] triton_kernel -> origin/triton_kernel 2025-09-07T07:36:04.2991406Z * [new branch] triton_kernel_perf -> origin/triton_kernel_perf 2025-09-07T07:36:04.2992662Z * [new branch] tt_pkg_1908 -> origin/tt_pkg_1908 2025-09-07T07:36:04.2994020Z * [new branch] tweak-transformer-dependabot -> origin/tweak-transformer-dependabot 2025-09-07T07:36:04.2995189Z * [new branch] type_dec -> origin/type_dec 2025-09-07T07:36:04.2996598Z * [new branch] udate-sphinx-dependancies -> origin/udate-sphinx-dependancies 2025-09-07T07:36:04.2998837Z * [new branch] update-audio-commit-hash/16818882925-1712-1 -> origin/update-audio-commit-hash/16818882925-1712-1 2025-09-07T07:36:04.3000070Z * [new branch] update-audio-commit-hash/16895560422-1720-1 -> origin/update-audio-commit-hash/16895560422-1720-1 2025-09-07T07:36:04.3001333Z * [new branch] update-audio-commit-hash/16924174496-1738-1 -> origin/update-audio-commit-hash/16924174496-1738-1 2025-09-07T07:36:04.3002576Z * [new branch] update-audio-commit-hash/17002010821-1749-1 -> origin/update-audio-commit-hash/17002010821-1749-1 2025-09-07T07:36:04.3007481Z * [new branch] update-audio-commit-hash/17056004427-1766-1 -> origin/update-audio-commit-hash/17056004427-1766-1 2025-09-07T07:36:04.3008949Z * [new branch] update-audio-commit-hash/17085054029-1767-1 -> origin/update-audio-commit-hash/17085054029-1767-1 2025-09-07T07:36:04.3010430Z * [new branch] update-audio-commit-hash/17142507405-1771-1 -> origin/update-audio-commit-hash/17142507405-1771-1 2025-09-07T07:36:04.3012008Z * [new branch] update-audio-commit-hash/17168762740-1773-1 -> origin/update-audio-commit-hash/17168762740-1773-1 2025-09-07T07:36:04.3013240Z * [new branch] update-audio-commit-hash/17311174639-1780-1 -> origin/update-audio-commit-hash/17311174639-1780-1 2025-09-07T07:36:04.3014426Z * [new branch] update-audio-commit-hash/17336898740-1781-1 -> origin/update-audio-commit-hash/17336898740-1781-1 2025-09-07T07:36:04.3015622Z * [new branch] update-audio-commit-hash/17389727684-1786-1 -> origin/update-audio-commit-hash/17389727684-1786-1 2025-09-07T07:36:04.3016812Z * [new branch] update-audio-commit-hash/17449538142-1790-1 -> origin/update-audio-commit-hash/17449538142-1790-1 2025-09-07T07:36:04.3018107Z * [new branch] update-audio-commit-hash/17507351808-1794-1 -> origin/update-audio-commit-hash/17507351808-1794-1 2025-09-07T07:36:04.3019344Z * [new branch] update-dynamic-shapes-doc -> origin/update-dynamic-shapes-doc 2025-09-07T07:36:04.3021180Z * [new branch] update-executorch-commit-hash/15694981040-1626-1 -> origin/update-executorch-commit-hash/15694981040-1626-1 2025-09-07T07:36:04.3022848Z * [new branch] update-triton-commit-hash/13663274526-1487-2 -> origin/update-triton-commit-hash/13663274526-1487-2 2025-09-07T07:36:04.3024541Z * [new branch] update-vision-commit-hash/15336342773-1607-1 -> origin/update-vision-commit-hash/15336342773-1607-1 2025-09-07T07:36:04.3026324Z * [new branch] update-vllm-commit-hash/16737365217-1704-1 -> origin/update-vllm-commit-hash/16737365217-1704-1 2025-09-07T07:36:04.3027469Z * [new branch] update-vllm-commit-hash/16843157111-1713-1 -> origin/update-vllm-commit-hash/16843157111-1713-1 2025-09-07T07:36:04.3028534Z * [new branch] update-vllm-commit-hash/16855312394-1714-1 -> origin/update-vllm-commit-hash/16855312394-1714-1 2025-09-07T07:36:04.3029660Z * [new branch] update-vllm-commit-hash/16924174496-1738-1 -> origin/update-vllm-commit-hash/16924174496-1738-1 2025-09-07T07:36:04.3030766Z * [new branch] update-vllm-commit-hash/16952608705-1745-1 -> origin/update-vllm-commit-hash/16952608705-1745-1 2025-09-07T07:36:04.3032163Z * [new branch] update-vllm-commit-hash/16979836546-1748-1 -> origin/update-vllm-commit-hash/16979836546-1748-1 2025-09-07T07:36:04.3033714Z * [new branch] update-vllm-commit-hash/17014576881-1756-1 -> origin/update-vllm-commit-hash/17014576881-1756-1 2025-09-07T07:36:04.3035207Z * [new branch] update-vllm-commit-hash/17027830869-1761-1 -> origin/update-vllm-commit-hash/17027830869-1761-1 2025-09-07T07:36:04.3036468Z * [new branch] update-vllm-commit-hash/17056004427-1766-1 -> origin/update-vllm-commit-hash/17056004427-1766-1 2025-09-07T07:36:04.3037676Z * [new branch] update-vllm-commit-hash/17085054029-1767-1 -> origin/update-vllm-commit-hash/17085054029-1767-1 2025-09-07T07:36:04.3038978Z * [new branch] update-vllm-commit-hash/17113610216-1768-1 -> origin/update-vllm-commit-hash/17113610216-1768-1 2025-09-07T07:36:04.3040680Z * [new branch] update-vllm-commit-hash/17142507405-1771-1 -> origin/update-vllm-commit-hash/17142507405-1771-1 2025-09-07T07:36:04.3041663Z * [new branch] update-vllm-commit-hash/17181878974-1774-1 -> origin/update-vllm-commit-hash/17181878974-1774-1 2025-09-07T07:36:04.3043133Z * [new branch] update-vllm-commit-hash/17311174639-1780-1 -> origin/update-vllm-commit-hash/17311174639-1780-1 2025-09-07T07:36:04.3044534Z * [new branch] update-vllm-commit-hash/17336898740-1781-1 -> origin/update-vllm-commit-hash/17336898740-1781-1 2025-09-07T07:36:04.3045833Z * [new branch] update-vllm-commit-hash/17364352302-1785-1 -> origin/update-vllm-commit-hash/17364352302-1785-1 2025-09-07T07:36:04.3047030Z * [new branch] update-vllm-commit-hash/17389727684-1786-1 -> origin/update-vllm-commit-hash/17389727684-1786-1 2025-09-07T07:36:04.3048291Z * [new branch] update-vllm-commit-hash/17449538142-1790-1 -> origin/update-vllm-commit-hash/17449538142-1790-1 2025-09-07T07:36:04.3049521Z * [new branch] update-vllm-commit-hash/17480069797-1791-1 -> origin/update-vllm-commit-hash/17480069797-1791-1 2025-09-07T07:36:04.3050638Z * [new branch] update-vllm-commit-hash/17507351808-1794-1 -> origin/update-vllm-commit-hash/17507351808-1794-1 2025-09-07T07:36:04.3052390Z * [new branch] update-xla-commit-hash/16873912760-198-1 -> origin/update-xla-commit-hash/16873912760-198-1 2025-09-07T07:36:04.3053562Z * [new branch] update-xla-commit-hash/17034266655-199-1 -> origin/update-xla-commit-hash/17034266655-199-1 2025-09-07T07:36:04.3054716Z * [new branch] update-xla-commit-hash/17202464405-200-1 -> origin/update-xla-commit-hash/17202464405-200-1 2025-09-07T07:36:04.3055973Z * [new branch] update_docs_torch_multinomial_issue#125388 -> origin/update_docs_torch_multinomial_issue#125388 2025-09-07T07:36:04.3057193Z * [new branch] update_executorch_pin -> origin/update_executorch_pin 2025-09-07T07:36:04.3058654Z * [new branch] update_slow_tests_1722488736 -> origin/update_slow_tests_1722488736 2025-09-07T07:36:04.3059989Z * [new branch] update_slow_tests_1722879173 -> origin/update_slow_tests_1722879173 2025-09-07T07:36:04.3061278Z * [new branch] update_slow_tests_1752478971 -> origin/update_slow_tests_1752478971 2025-09-07T07:36:04.3062522Z * [new branch] update_slow_tests_1755502951 -> origin/update_slow_tests_1755502951 2025-09-07T07:36:04.3063786Z * [new branch] update_slow_tests_1756107664 -> origin/update_slow_tests_1756107664 2025-09-07T07:36:04.3065573Z * [new branch] update_submodule_FBGEMM -> origin/update_submodule_FBGEMM 2025-09-07T07:36:04.3066889Z * [new branch] update_submodule_kineto -> origin/update_submodule_kineto 2025-09-07T07:36:04.3068178Z * [new branch] update_submodule_tensorpipe -> origin/update_submodule_tensorpipe 2025-09-07T07:36:04.3069584Z * [new branch] v0.1.2 -> origin/v0.1.2 2025-09-07T07:36:04.3070976Z * [new branch] v1.0.1 -> origin/v1.0.1 2025-09-07T07:36:04.3072341Z * [new branch] v1.0.3 -> origin/v1.0.3 2025-09-07T07:36:04.3073833Z * [new branch] v1.1.0 -> origin/v1.1.0 2025-09-07T07:36:04.3075203Z * [new branch] v1.2.0 -> origin/v1.2.0 2025-09-07T07:36:04.3076555Z * [new branch] v1.3.0 -> origin/v1.3.0 2025-09-07T07:36:04.3077907Z * [new branch] v1.3.1 -> origin/v1.3.1 2025-09-07T07:36:04.3079238Z * [new branch] validate_fn -> origin/validate_fn 2025-09-07T07:36:04.3080663Z * [new branch] validations_2.6 -> origin/validations_2.6 2025-09-07T07:36:04.3082118Z * [new branch] validations_2.8 -> origin/validations_2.8 2025-09-07T07:36:04.3084021Z * [new branch] viable/strict -> origin/viable/strict 2025-09-07T07:36:04.3085221Z * [new branch] vllmbuildci -> origin/vllmbuildci 2025-09-07T07:36:04.3086629Z * [new branch] vllmpin -> origin/vllmpin 2025-09-07T07:36:04.3088819Z * [new branch] wdvr/conda_devcontainer -> origin/wdvr/conda_devcontainer 2025-09-07T07:36:04.3089952Z * [new branch] wdvr/iss_145259 -> origin/wdvr/iss_145259 2025-09-07T07:36:04.3091394Z * [new branch] weight_sharing_cpp -> origin/weight_sharing_cpp 2025-09-07T07:36:04.3093461Z * [new branch] whc/flight4 -> origin/whc/flight4 2025-09-07T07:36:04.3094613Z * [new branch] whc/flight51 -> origin/whc/flight51 2025-09-07T07:36:04.3095777Z * [new branch] whc/flight53 -> origin/whc/flight53 2025-09-07T07:36:04.3097060Z * [new branch] whc/stage2 -> origin/whc/stage2 2025-09-07T07:36:04.3098220Z * [new branch] whc/uneven -> origin/whc/uneven 2025-09-07T07:36:04.3099667Z * [new branch] whc/uneven-merge -> origin/whc/uneven-merge 2025-09-07T07:36:04.3100970Z * [new branch] win_warnings -> origin/win_warnings 2025-09-07T07:36:04.3102225Z * [new branch] windows_libtorch_free -> origin/windows_libtorch_free 2025-09-07T07:36:04.3103587Z * [new branch] workonoldcommit -> origin/workonoldcommit 2025-09-07T07:36:04.3105135Z * [new branch] wychi-autotune-prune-configs-by-shared-mem -> origin/wychi-autotune-prune-configs-by-shared-mem 2025-09-07T07:36:04.3106804Z * [new branch] xmfan/ca_0516 -> origin/xmfan/ca_0516 2025-09-07T07:36:04.3107943Z * [new branch] xmfan/ca_1051b93192 -> origin/xmfan/ca_1051b93192 2025-09-07T07:36:04.3109314Z * [new branch] xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 -> origin/xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 2025-09-07T07:36:04.3110275Z * [new branch] xmfan/ca_5a2be192d1 -> origin/xmfan/ca_5a2be192d1 2025-09-07T07:36:04.3111360Z * [new branch] xmfan/ca_9d59b516e9 -> origin/xmfan/ca_9d59b516e9 2025-09-07T07:36:04.3112535Z * [new branch] xmfan/ca_api -> origin/xmfan/ca_api 2025-09-07T07:36:04.3113655Z * [new branch] xmfan/ca_apr8 -> origin/xmfan/ca_apr8 2025-09-07T07:36:04.3115051Z * [new branch] xmfan/ca_base -> origin/xmfan/ca_base 2025-09-07T07:36:04.3116637Z * [new branch] xmfan/ca_cudagraphs -> origin/xmfan/ca_cudagraphs 2025-09-07T07:36:04.3117834Z * [new branch] xmfan/ca_dynamic -> origin/xmfan/ca_dynamic 2025-09-07T07:36:04.3119082Z * [new branch] xmfan/ca_fix_dyn -> origin/xmfan/ca_fix_dyn 2025-09-07T07:36:04.3120275Z * [new branch] xmfan/ca_fix_lowering -> origin/xmfan/ca_fix_lowering 2025-09-07T07:36:04.3121559Z * [new branch] xmfan/ca_fix_polyfills -> origin/xmfan/ca_fix_polyfills 2025-09-07T07:36:04.3122646Z * [new branch] xmfan/ca_jan3 -> origin/xmfan/ca_jan3 2025-09-07T07:36:04.3124274Z * [new branch] xmfan/ca_jun18 -> origin/xmfan/ca_jun18 2025-09-07T07:36:04.3125531Z * [new branch] xmfan/ca_jun24 -> origin/xmfan/ca_jun24 2025-09-07T07:36:04.3126721Z * [new branch] xmfan/ca_mem_base -> origin/xmfan/ca_mem_base 2025-09-07T07:36:04.3127889Z * [new branch] xmfan/ca_mem_fix -> origin/xmfan/ca_mem_fix 2025-09-07T07:36:04.3129118Z * [new branch] xmfan/ca_memory_fix -> origin/xmfan/ca_memory_fix 2025-09-07T07:36:04.3130299Z * [new branch] xmfan/ca_memory_fix_rebased -> origin/xmfan/ca_memory_fix_rebased 2025-09-07T07:36:04.3131614Z * [new branch] xmfan/ca_memory_fix_rebased2 -> origin/xmfan/ca_memory_fix_rebased2 2025-09-07T07:36:04.3132751Z * [new branch] xmfan/ca_move_to_cuda -> origin/xmfan/ca_move_to_cuda 2025-09-07T07:36:04.3133941Z * [new branch] xmfan/ca_nested -> origin/xmfan/ca_nested 2025-09-07T07:36:04.3135161Z * [new branch] xmfan/ca_overhead -> origin/xmfan/ca_overhead 2025-09-07T07:36:04.3136378Z * [new branch] xmfan/ca_overhead_0eba7e5451 -> origin/xmfan/ca_overhead_0eba7e5451 2025-09-07T07:36:04.3137524Z * [new branch] xmfan/ca_scalar -> origin/xmfan/ca_scalar 2025-09-07T07:36:04.3138839Z * [new branch] xmfan/ca_subclass_mem_fix -> origin/xmfan/ca_subclass_mem_fix 2025-09-07T07:36:04.3140088Z * [new branch] xmfan/ca_warm_mem -> origin/xmfan/ca_warm_mem 2025-09-07T07:36:04.3141496Z * [new branch] xmfan/ca_warm_mem_base -> origin/xmfan/ca_warm_mem_base 2025-09-07T07:36:04.3142738Z * [new branch] xmfan/cacu_jun18 -> origin/xmfan/cacu_jun18 2025-09-07T07:36:04.3143879Z * [new branch] xmfan/cacu_jun19 -> origin/xmfan/cacu_jun19 2025-09-07T07:36:04.3145154Z * [new branch] xmfan/cacu_jun4 -> origin/xmfan/cacu_jun4 2025-09-07T07:36:04.3146536Z * [new branch] xmfan/cacu_may27 -> origin/xmfan/cacu_may27 2025-09-07T07:36:04.3147902Z * [new branch] xmfan/disable_duck_shape -> origin/xmfan/disable_duck_shape 2025-09-07T07:36:04.3149312Z * [new branch] xmfan/fca_cpp_node_passthrough -> origin/xmfan/fca_cpp_node_passthrough 2025-09-07T07:36:04.3150590Z * [new branch] xmfan/issue_123374 -> origin/xmfan/issue_123374 2025-09-07T07:36:04.3152001Z * [new branch] xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 2025-09-07T07:36:04.3153698Z * [new branch] xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 2025-09-07T07:36:04.3154812Z * [new branch] xmfan/segfault_test -> origin/xmfan/segfault_test 2025-09-07T07:36:04.3156022Z * [new branch] xmfan/single_step -> origin/xmfan/single_step 2025-09-07T07:36:04.3157232Z * [new branch] xmfan/sth_0829 -> origin/xmfan/sth_0829 2025-09-07T07:36:04.3158649Z * [new branch] xmfan/test -> origin/xmfan/test 2025-09-07T07:36:04.3160501Z * [new branch] yguo/debug-0226-constexpr -> origin/yguo/debug-0226-constexpr 2025-09-07T07:36:04.3161622Z * [new branch] yguo/new_latest_changes -> origin/yguo/new_latest_changes 2025-09-07T07:36:04.3162753Z * [new branch] yguo/patch_constexpr_changes -> origin/yguo/patch_constexpr_changes 2025-09-07T07:36:04.3164010Z * [new branch] yihan_quantization -> origin/yihan_quantization 2025-09-07T07:36:04.3166239Z * [new branch] yiming/add_jit_trace_benchmark -> origin/yiming/add_jit_trace_benchmark 2025-09-07T07:36:04.3167406Z * [new branch] yiming/add_nativert_benchmark -> origin/yiming/add_nativert_benchmark 2025-09-07T07:36:04.3168506Z * [new branch] yiming/bootcamp -> origin/yiming/bootcamp 2025-09-07T07:36:04.3170276Z * [new branch] zainr/canary-test -> origin/zainr/canary-test 2025-09-07T07:36:04.3171931Z * [new branch] zainr/cleanup-gh-runners -> origin/zainr/cleanup-gh-runners 2025-09-07T07:36:04.3173049Z * [new branch] zainr/git-push-v2 -> origin/zainr/git-push-v2 2025-09-07T07:36:04.3174236Z * [new branch] zainr/pull-migration-c -> origin/zainr/pull-migration-c 2025-09-07T07:36:04.3175523Z * [new branch] zainr/test -> origin/zainr/test 2025-09-07T07:36:04.3176705Z * [new branch] zainr/test2 -> origin/zainr/test2 2025-09-07T07:36:04.3177789Z * [new branch] zainr/unstable -> origin/zainr/unstable 2025-09-07T07:36:04.3178941Z * [new branch] zainr/unstable-xla -> origin/zainr/unstable-xla 2025-09-07T07:36:04.3180322Z * [new branch] zasdfgbnm-patch-3 -> origin/zasdfgbnm-patch-3 2025-09-07T07:36:04.3181608Z * [new branch] zb2p -> origin/zb2p 2025-09-07T07:36:04.3183315Z * [new branch] zero_grad_optimization -> origin/zero_grad_optimization 2025-09-07T07:36:04.3184699Z * [new branch] zeros-and-scatter-part2 -> origin/zeros-and-scatter-part2 2025-09-07T07:36:04.3186822Z * [new branch] zhxchen17/scratch/0 -> origin/zhxchen17/scratch/0 2025-09-07T07:36:04.3188671Z * [new branch] zhxhcen17/moodycamel -> origin/zhxhcen17/moodycamel 2025-09-07T07:36:04.3190381Z * [new branch] zxiiro/main -> origin/zxiiro/main 2025-09-07T07:36:04.3191642Z * [new tag] bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug -> bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug 2025-09-07T07:36:04.3192698Z * [new tag] ci/binaries/77164 -> ci/binaries/77164 2025-09-07T07:36:04.3193841Z * [new tag] ciflow/binaries/156049 -> ciflow/binaries/156049 2025-09-07T07:36:04.3194553Z * [new tag] ciflow/binaries/156712 -> ciflow/binaries/156712 2025-09-07T07:36:04.3195382Z * [new tag] ciflow/binaries/157432 -> ciflow/binaries/157432 2025-09-07T07:36:04.3196206Z * [new tag] ciflow/binaries/157685 -> ciflow/binaries/157685 2025-09-07T07:36:04.3196926Z * [new tag] ciflow/binaries/157689 -> ciflow/binaries/157689 2025-09-07T07:36:04.3197751Z * [new tag] ciflow/binaries/158104 -> ciflow/binaries/158104 2025-09-07T07:36:04.3198619Z * [new tag] ciflow/binaries/160229 -> ciflow/binaries/160229 2025-09-07T07:36:04.3199502Z * [new tag] ciflow/binaries/160720 -> ciflow/binaries/160720 2025-09-07T07:36:04.3200372Z * [new tag] ciflow/binaries/162080 -> ciflow/binaries/162080 2025-09-07T07:36:04.3201022Z * [new tag] ciflow/binaries/162329 -> ciflow/binaries/162329 2025-09-07T07:36:04.3202152Z * [new tag] ciflow/binaries_libtorch/156049 -> ciflow/binaries_libtorch/156049 2025-09-07T07:36:04.3203021Z * [new tag] ciflow/binaries_libtorch/156711 -> ciflow/binaries_libtorch/156711 2025-09-07T07:36:04.3204018Z * [new tag] ciflow/binaries_libtorch/157432 -> ciflow/binaries_libtorch/157432 2025-09-07T07:36:04.3204972Z * [new tag] ciflow/binaries_wheel/156049 -> ciflow/binaries_wheel/156049 2025-09-07T07:36:04.3205721Z * [new tag] ciflow/binaries_wheel/156711 -> ciflow/binaries_wheel/156711 2025-09-07T07:36:04.3207115Z * [new tag] ciflow/binaries_wheel/157432 -> ciflow/binaries_wheel/157432 2025-09-07T07:36:04.3207400Z * [new tag] ciflow/binaries_wheel/162136 -> ciflow/binaries_wheel/162136 2025-09-07T07:36:04.3208123Z * [new tag] ciflow/binaries_wheel/162252 -> ciflow/binaries_wheel/162252 2025-09-07T07:36:04.3209022Z * [new tag] ciflow/binaries_wheel/162325 -> ciflow/binaries_wheel/162325 2025-09-07T07:36:04.3210026Z * [new tag] ciflow/h100-distributed/156703 -> ciflow/h100-distributed/156703 2025-09-07T07:36:04.3210969Z * [new tag] ciflow/h100-symm-mem/157635 -> ciflow/h100-symm-mem/157635 2025-09-07T07:36:04.3211684Z * [new tag] ciflow/h100-symm-mem/161984 -> ciflow/h100-symm-mem/161984 2025-09-07T07:36:04.3212550Z * [new tag] ciflow/h100-symm-mem/162003 -> ciflow/h100-symm-mem/162003 2025-09-07T07:36:04.3213662Z * [new tag] ciflow/h100-symm-mem/162011 -> ciflow/h100-symm-mem/162011 2025-09-07T07:36:04.3214034Z * [new tag] ciflow/h100-symm-mem/162026 -> ciflow/h100-symm-mem/162026 2025-09-07T07:36:04.3215543Z * [new tag] ciflow/h100-symm-mem/162033 -> ciflow/h100-symm-mem/162033 2025-09-07T07:36:04.3216365Z * [new tag] ciflow/h100-symm-mem/162040 -> ciflow/h100-symm-mem/162040 2025-09-07T07:36:04.3217083Z * [new tag] ciflow/h100-symm-mem/162041 -> ciflow/h100-symm-mem/162041 2025-09-07T07:36:04.3217992Z * [new tag] ciflow/h100-symm-mem/162142 -> ciflow/h100-symm-mem/162142 2025-09-07T07:36:04.3218776Z * [new tag] ciflow/h100-symm-mem/162150 -> ciflow/h100-symm-mem/162150 2025-09-07T07:36:04.3219595Z * [new tag] ciflow/h100-symm-mem/162243 -> ciflow/h100-symm-mem/162243 2025-09-07T07:36:04.3220637Z * [new tag] ciflow/h100-symm-mem/162320 -> ciflow/h100-symm-mem/162320 2025-09-07T07:36:04.3221638Z * [new tag] ciflow/h100/159158 -> ciflow/h100/159158 2025-09-07T07:36:04.3222831Z * [new tag] ciflow/h100/160480 -> ciflow/h100/160480 2025-09-07T07:36:04.3223695Z * [new tag] ciflow/h100/161749 -> ciflow/h100/161749 2025-09-07T07:36:04.3224606Z * [new tag] ciflow/h100/162022 -> ciflow/h100/162022 2025-09-07T07:36:04.3225449Z * [new tag] ciflow/h100/162278 -> ciflow/h100/162278 2025-09-07T07:36:04.3226671Z * [new tag] ciflow/inductor-perf-test-nightly-rocm/156592 -> ciflow/inductor-perf-test-nightly-rocm/156592 2025-09-07T07:36:04.3227764Z * [new tag] ciflow/inductor-perf-test-nightly/156592 -> ciflow/inductor-perf-test-nightly/156592 2025-09-07T07:36:04.3228814Z * [new tag] ciflow/inductor-periodic/162063 -> ciflow/inductor-periodic/162063 2025-09-07T07:36:04.3229516Z * [new tag] ciflow/inductor-periodic/162227 -> ciflow/inductor-periodic/162227 2025-09-07T07:36:04.3230442Z * [new tag] ciflow/inductor-periodic/162323 -> ciflow/inductor-periodic/162323 2025-09-07T07:36:04.3231488Z * [new tag] ciflow/inductor-rocm/154170 -> ciflow/inductor-rocm/154170 2025-09-07T07:36:04.3232471Z * [new tag] ciflow/inductor-rocm/159146 -> ciflow/inductor-rocm/159146 2025-09-07T07:36:04.3233172Z * [new tag] ciflow/inductor-rocm/159158 -> ciflow/inductor-rocm/159158 2025-09-07T07:36:04.3234139Z * [new tag] ciflow/inductor-rocm/161715 -> ciflow/inductor-rocm/161715 2025-09-07T07:36:04.3235064Z * [new tag] ciflow/inductor-rocm/162053 -> ciflow/inductor-rocm/162053 2025-09-07T07:36:04.3236003Z * [new tag] ciflow/inductor-rocm/162056 -> ciflow/inductor-rocm/162056 2025-09-07T07:36:04.3237041Z * [new tag] ciflow/inductor/137400 -> ciflow/inductor/137400 2025-09-07T07:36:04.3237831Z * [new tag] ciflow/inductor/148180 -> ciflow/inductor/148180 2025-09-07T07:36:04.3238538Z * [new tag] ciflow/inductor/148328 -> ciflow/inductor/148328 2025-09-07T07:36:04.3239347Z * [new tag] ciflow/inductor/148484 -> ciflow/inductor/148484 2025-09-07T07:36:04.3240177Z * [new tag] ciflow/inductor/148492 -> ciflow/inductor/148492 2025-09-07T07:36:04.3240832Z * [new tag] ciflow/inductor/152624 -> ciflow/inductor/152624 2025-09-07T07:36:04.3241670Z * [new tag] ciflow/inductor/154694 -> ciflow/inductor/154694 2025-09-07T07:36:04.3242359Z * [new tag] ciflow/inductor/156049 -> ciflow/inductor/156049 2025-09-07T07:36:04.3243201Z * [new tag] ciflow/inductor/156592 -> ciflow/inductor/156592 2025-09-07T07:36:04.3244040Z * [new tag] ciflow/inductor/157635 -> ciflow/inductor/157635 2025-09-07T07:36:04.3244767Z * [new tag] ciflow/inductor/157685 -> ciflow/inductor/157685 2025-09-07T07:36:04.3245795Z * [new tag] ciflow/inductor/157686 -> ciflow/inductor/157686 2025-09-07T07:36:04.3246839Z * [new tag] ciflow/inductor/157689 -> ciflow/inductor/157689 2025-09-07T07:36:04.3247851Z * [new tag] ciflow/inductor/157699 -> ciflow/inductor/157699 2025-09-07T07:36:04.3248788Z * [new tag] ciflow/inductor/157743 -> ciflow/inductor/157743 2025-09-07T07:36:04.3249733Z * [new tag] ciflow/inductor/157994 -> ciflow/inductor/157994 2025-09-07T07:36:04.3250557Z * [new tag] ciflow/inductor/158091 -> ciflow/inductor/158091 2025-09-07T07:36:04.3251470Z * [new tag] ciflow/inductor/158104 -> ciflow/inductor/158104 2025-09-07T07:36:04.3252404Z * [new tag] ciflow/inductor/158404 -> ciflow/inductor/158404 2025-09-07T07:36:04.3253255Z * [new tag] ciflow/inductor/158647 -> ciflow/inductor/158647 2025-09-07T07:36:04.3254253Z * [new tag] ciflow/inductor/158932 -> ciflow/inductor/158932 2025-09-07T07:36:04.3255187Z * [new tag] ciflow/inductor/159146 -> ciflow/inductor/159146 2025-09-07T07:36:04.3256014Z * [new tag] ciflow/inductor/159158 -> ciflow/inductor/159158 2025-09-07T07:36:04.3256947Z * [new tag] ciflow/inductor/159274 -> ciflow/inductor/159274 2025-09-07T07:36:04.3257801Z * [new tag] ciflow/inductor/159664 -> ciflow/inductor/159664 2025-09-07T07:36:04.3258896Z * [new tag] ciflow/inductor/159778 -> ciflow/inductor/159778 2025-09-07T07:36:04.3259720Z * [new tag] ciflow/inductor/159835 -> ciflow/inductor/159835 2025-09-07T07:36:04.3260889Z * [new tag] ciflow/inductor/159944 -> ciflow/inductor/159944 2025-09-07T07:36:04.3261907Z * [new tag] ciflow/inductor/160161 -> ciflow/inductor/160161 2025-09-07T07:36:04.3262730Z * [new tag] ciflow/inductor/160174 -> ciflow/inductor/160174 2025-09-07T07:36:04.3263681Z * [new tag] ciflow/inductor/160323 -> ciflow/inductor/160323 2025-09-07T07:36:04.3264765Z * [new tag] ciflow/inductor/160324 -> ciflow/inductor/160324 2025-09-07T07:36:04.3265721Z * [new tag] ciflow/inductor/160325 -> ciflow/inductor/160325 2025-09-07T07:36:04.3266732Z * [new tag] ciflow/inductor/160326 -> ciflow/inductor/160326 2025-09-07T07:36:04.3267572Z * [new tag] ciflow/inductor/160327 -> ciflow/inductor/160327 2025-09-07T07:36:04.3268456Z * [new tag] ciflow/inductor/160328 -> ciflow/inductor/160328 2025-09-07T07:36:04.3269397Z * [new tag] ciflow/inductor/160329 -> ciflow/inductor/160329 2025-09-07T07:36:04.3270241Z * [new tag] ciflow/inductor/160480 -> ciflow/inductor/160480 2025-09-07T07:36:04.3271215Z * [new tag] ciflow/inductor/160532 -> ciflow/inductor/160532 2025-09-07T07:36:04.3272583Z * [new tag] ciflow/inductor/160539 -> ciflow/inductor/160539 2025-09-07T07:36:04.3273450Z * [new tag] ciflow/inductor/160580 -> ciflow/inductor/160580 2025-09-07T07:36:04.3274306Z * [new tag] ciflow/inductor/160685 -> ciflow/inductor/160685 2025-09-07T07:36:04.3275123Z * [new tag] ciflow/inductor/160686 -> ciflow/inductor/160686 2025-09-07T07:36:04.3275962Z * [new tag] ciflow/inductor/160687 -> ciflow/inductor/160687 2025-09-07T07:36:04.3276803Z * [new tag] ciflow/inductor/160688 -> ciflow/inductor/160688 2025-09-07T07:36:04.3277643Z * [new tag] ciflow/inductor/160690 -> ciflow/inductor/160690 2025-09-07T07:36:04.3278465Z * [new tag] ciflow/inductor/160706 -> ciflow/inductor/160706 2025-09-07T07:36:04.3279380Z * [new tag] ciflow/inductor/160729 -> ciflow/inductor/160729 2025-09-07T07:36:04.3280220Z * [new tag] ciflow/inductor/160798 -> ciflow/inductor/160798 2025-09-07T07:36:04.3281230Z * [new tag] ciflow/inductor/160836 -> ciflow/inductor/160836 2025-09-07T07:36:04.3282068Z * [new tag] ciflow/inductor/160843 -> ciflow/inductor/160843 2025-09-07T07:36:04.3283193Z * [new tag] ciflow/inductor/160869 -> ciflow/inductor/160869 2025-09-07T07:36:04.3284012Z * [new tag] ciflow/inductor/160920 -> ciflow/inductor/160920 2025-09-07T07:36:04.3284874Z * [new tag] ciflow/inductor/160943 -> ciflow/inductor/160943 2025-09-07T07:36:04.3285704Z * [new tag] ciflow/inductor/161092 -> ciflow/inductor/161092 2025-09-07T07:36:04.3287004Z * [new tag] ciflow/inductor/161093 -> ciflow/inductor/161093 2025-09-07T07:36:04.3287946Z * [new tag] ciflow/inductor/161109 -> ciflow/inductor/161109 2025-09-07T07:36:04.3288775Z * [new tag] ciflow/inductor/161118 -> ciflow/inductor/161118 2025-09-07T07:36:04.3289737Z * [new tag] ciflow/inductor/161178 -> ciflow/inductor/161178 2025-09-07T07:36:04.3290665Z * [new tag] ciflow/inductor/161246 -> ciflow/inductor/161246 2025-09-07T07:36:04.3291491Z * [new tag] ciflow/inductor/161349 -> ciflow/inductor/161349 2025-09-07T07:36:04.3292350Z * [new tag] ciflow/inductor/161350 -> ciflow/inductor/161350 2025-09-07T07:36:04.3293178Z * [new tag] ciflow/inductor/161351 -> ciflow/inductor/161351 2025-09-07T07:36:04.3294146Z * [new tag] ciflow/inductor/161397 -> ciflow/inductor/161397 2025-09-07T07:36:04.3295131Z * [new tag] ciflow/inductor/161404 -> ciflow/inductor/161404 2025-09-07T07:36:04.3295767Z * [new tag] ciflow/inductor/161405 -> ciflow/inductor/161405 2025-09-07T07:36:04.3296693Z * [new tag] ciflow/inductor/161406 -> ciflow/inductor/161406 2025-09-07T07:36:04.3297740Z * [new tag] ciflow/inductor/161410 -> ciflow/inductor/161410 2025-09-07T07:36:04.3298627Z * [new tag] ciflow/inductor/161414 -> ciflow/inductor/161414 2025-09-07T07:36:04.3299654Z * [new tag] ciflow/inductor/161442 -> ciflow/inductor/161442 2025-09-07T07:36:04.3300499Z * [new tag] ciflow/inductor/161458 -> ciflow/inductor/161458 2025-09-07T07:36:04.3301311Z * [new tag] ciflow/inductor/161468 -> ciflow/inductor/161468 2025-09-07T07:36:04.3302171Z * [new tag] ciflow/inductor/161469 -> ciflow/inductor/161469 2025-09-07T07:36:04.3303240Z * [new tag] ciflow/inductor/161485 -> ciflow/inductor/161485 2025-09-07T07:36:04.3304224Z * [new tag] ciflow/inductor/161499 -> ciflow/inductor/161499 2025-09-07T07:36:04.3305044Z * [new tag] ciflow/inductor/161534 -> ciflow/inductor/161534 2025-09-07T07:36:04.3306017Z * [new tag] ciflow/inductor/161595 -> ciflow/inductor/161595 2025-09-07T07:36:04.3306859Z * [new tag] ciflow/inductor/161596 -> ciflow/inductor/161596 2025-09-07T07:36:04.3308099Z * [new tag] ciflow/inductor/161630 -> ciflow/inductor/161630 2025-09-07T07:36:04.3308923Z * [new tag] ciflow/inductor/161667 -> ciflow/inductor/161667 2025-09-07T07:36:04.3309742Z * [new tag] ciflow/inductor/161670 -> ciflow/inductor/161670 2025-09-07T07:36:04.3310662Z * [new tag] ciflow/inductor/161673 -> ciflow/inductor/161673 2025-09-07T07:36:04.3311472Z * [new tag] ciflow/inductor/161674 -> ciflow/inductor/161674 2025-09-07T07:36:04.3312511Z * [new tag] ciflow/inductor/161675 -> ciflow/inductor/161675 2025-09-07T07:36:04.3313084Z * [new tag] ciflow/inductor/161693 -> ciflow/inductor/161693 2025-09-07T07:36:04.3313964Z * [new tag] ciflow/inductor/161695 -> ciflow/inductor/161695 2025-09-07T07:36:04.3314790Z * [new tag] ciflow/inductor/161715 -> ciflow/inductor/161715 2025-09-07T07:36:04.3315630Z * [new tag] ciflow/inductor/161730 -> ciflow/inductor/161730 2025-09-07T07:36:04.3316470Z * [new tag] ciflow/inductor/161732 -> ciflow/inductor/161732 2025-09-07T07:36:04.3317383Z * [new tag] ciflow/inductor/161744 -> ciflow/inductor/161744 2025-09-07T07:36:04.3318242Z * [new tag] ciflow/inductor/161746 -> ciflow/inductor/161746 2025-09-07T07:36:04.3319040Z * [new tag] ciflow/inductor/161747 -> ciflow/inductor/161747 2025-09-07T07:36:04.3319918Z * [new tag] ciflow/inductor/161819 -> ciflow/inductor/161819 2025-09-07T07:36:04.3320802Z * [new tag] ciflow/inductor/161821 -> ciflow/inductor/161821 2025-09-07T07:36:04.3321513Z * [new tag] ciflow/inductor/161828 -> ciflow/inductor/161828 2025-09-07T07:36:04.3322306Z * [new tag] ciflow/inductor/161879 -> ciflow/inductor/161879 2025-09-07T07:36:04.3323148Z * [new tag] ciflow/inductor/161880 -> ciflow/inductor/161880 2025-09-07T07:36:04.3323969Z * [new tag] ciflow/inductor/161881 -> ciflow/inductor/161881 2025-09-07T07:36:04.3325045Z * [new tag] ciflow/inductor/161907 -> ciflow/inductor/161907 2025-09-07T07:36:04.3325809Z * [new tag] ciflow/inductor/161914 -> ciflow/inductor/161914 2025-09-07T07:36:04.3326768Z * [new tag] ciflow/inductor/161924 -> ciflow/inductor/161924 2025-09-07T07:36:04.3327650Z * [new tag] ciflow/inductor/161936 -> ciflow/inductor/161936 2025-09-07T07:36:04.3328501Z * [new tag] ciflow/inductor/161938 -> ciflow/inductor/161938 2025-09-07T07:36:04.3329385Z * [new tag] ciflow/inductor/161939 -> ciflow/inductor/161939 2025-09-07T07:36:04.3330252Z * [new tag] ciflow/inductor/161940 -> ciflow/inductor/161940 2025-09-07T07:36:04.3331078Z * [new tag] ciflow/inductor/161955 -> ciflow/inductor/161955 2025-09-07T07:36:04.3331913Z * [new tag] ciflow/inductor/161957 -> ciflow/inductor/161957 2025-09-07T07:36:04.3332920Z * [new tag] ciflow/inductor/161975 -> ciflow/inductor/161975 2025-09-07T07:36:04.3333662Z * [new tag] ciflow/inductor/161977 -> ciflow/inductor/161977 2025-09-07T07:36:04.3334421Z * [new tag] ciflow/inductor/161978 -> ciflow/inductor/161978 2025-09-07T07:36:04.3335248Z * [new tag] ciflow/inductor/161979 -> ciflow/inductor/161979 2025-09-07T07:36:04.3336066Z * [new tag] ciflow/inductor/161980 -> ciflow/inductor/161980 2025-09-07T07:36:04.3337035Z * [new tag] ciflow/inductor/161988 -> ciflow/inductor/161988 2025-09-07T07:36:04.3337843Z * [new tag] ciflow/inductor/161994 -> ciflow/inductor/161994 2025-09-07T07:36:04.3338691Z * [new tag] ciflow/inductor/162013 -> ciflow/inductor/162013 2025-09-07T07:36:04.3339522Z * [new tag] ciflow/inductor/162014 -> ciflow/inductor/162014 2025-09-07T07:36:04.3340369Z * [new tag] ciflow/inductor/162017 -> ciflow/inductor/162017 2025-09-07T07:36:04.3341266Z * [new tag] ciflow/inductor/162021 -> ciflow/inductor/162021 2025-09-07T07:36:04.3342094Z * [new tag] ciflow/inductor/162023 -> ciflow/inductor/162023 2025-09-07T07:36:04.3343335Z * [new tag] ciflow/inductor/162027 -> ciflow/inductor/162027 2025-09-07T07:36:04.3344180Z * [new tag] ciflow/inductor/162029 -> ciflow/inductor/162029 2025-09-07T07:36:04.3344984Z * [new tag] ciflow/inductor/162030 -> ciflow/inductor/162030 2025-09-07T07:36:04.3345867Z * [new tag] ciflow/inductor/162031 -> ciflow/inductor/162031 2025-09-07T07:36:04.3346636Z * [new tag] ciflow/inductor/162033 -> ciflow/inductor/162033 2025-09-07T07:36:04.3347691Z * [new tag] ciflow/inductor/162052 -> ciflow/inductor/162052 2025-09-07T07:36:04.3348421Z * [new tag] ciflow/inductor/162053 -> ciflow/inductor/162053 2025-09-07T07:36:04.3349606Z * [new tag] ciflow/inductor/162056 -> ciflow/inductor/162056 2025-09-07T07:36:04.3350347Z * [new tag] ciflow/inductor/162063 -> ciflow/inductor/162063 2025-09-07T07:36:04.3351163Z * [new tag] ciflow/inductor/162066 -> ciflow/inductor/162066 2025-09-07T07:36:04.3351992Z * [new tag] ciflow/inductor/162068 -> ciflow/inductor/162068 2025-09-07T07:36:04.3353158Z * [new tag] ciflow/inductor/162081 -> ciflow/inductor/162081 2025-09-07T07:36:04.3353827Z * [new tag] ciflow/inductor/162088 -> ciflow/inductor/162088 2025-09-07T07:36:04.3354657Z * [new tag] ciflow/inductor/162089 -> ciflow/inductor/162089 2025-09-07T07:36:04.3355477Z * [new tag] ciflow/inductor/162094 -> ciflow/inductor/162094 2025-09-07T07:36:04.3356445Z * [new tag] ciflow/inductor/162098 -> ciflow/inductor/162098 2025-09-07T07:36:04.3357202Z * [new tag] ciflow/inductor/162101 -> ciflow/inductor/162101 2025-09-07T07:36:04.3358063Z * [new tag] ciflow/inductor/162102 -> ciflow/inductor/162102 2025-09-07T07:36:04.3358903Z * [new tag] ciflow/inductor/162104 -> ciflow/inductor/162104 2025-09-07T07:36:04.3359735Z * [new tag] ciflow/inductor/162106 -> ciflow/inductor/162106 2025-09-07T07:36:04.3360566Z * [new tag] ciflow/inductor/162108 -> ciflow/inductor/162108 2025-09-07T07:36:04.3361400Z * [new tag] ciflow/inductor/162126 -> ciflow/inductor/162126 2025-09-07T07:36:04.3362265Z * [new tag] ciflow/inductor/162149 -> ciflow/inductor/162149 2025-09-07T07:36:04.3363079Z * [new tag] ciflow/inductor/162164 -> ciflow/inductor/162164 2025-09-07T07:36:04.3363908Z * [new tag] ciflow/inductor/162166 -> ciflow/inductor/162166 2025-09-07T07:36:04.3364750Z * [new tag] ciflow/inductor/162169 -> ciflow/inductor/162169 2025-09-07T07:36:04.3365703Z * [new tag] ciflow/inductor/162170 -> ciflow/inductor/162170 2025-09-07T07:36:04.3366471Z * [new tag] ciflow/inductor/162171 -> ciflow/inductor/162171 2025-09-07T07:36:04.3367316Z * [new tag] ciflow/inductor/162183 -> ciflow/inductor/162183 2025-09-07T07:36:04.3368125Z * [new tag] ciflow/inductor/162189 -> ciflow/inductor/162189 2025-09-07T07:36:04.3368953Z * [new tag] ciflow/inductor/162190 -> ciflow/inductor/162190 2025-09-07T07:36:04.3369820Z * [new tag] ciflow/inductor/162191 -> ciflow/inductor/162191 2025-09-07T07:36:04.3370666Z * [new tag] ciflow/inductor/162194 -> ciflow/inductor/162194 2025-09-07T07:36:04.3371775Z * [new tag] ciflow/inductor/162200 -> ciflow/inductor/162200 2025-09-07T07:36:04.3372556Z * [new tag] ciflow/inductor/162201 -> ciflow/inductor/162201 2025-09-07T07:36:04.3373380Z * [new tag] ciflow/inductor/162208 -> ciflow/inductor/162208 2025-09-07T07:36:04.3374450Z * [new tag] ciflow/inductor/162211 -> ciflow/inductor/162211 2025-09-07T07:36:04.3375277Z * [new tag] ciflow/inductor/162216 -> ciflow/inductor/162216 2025-09-07T07:36:04.3376033Z * [new tag] ciflow/inductor/162220 -> ciflow/inductor/162220 2025-09-07T07:36:04.3377064Z * [new tag] ciflow/inductor/162222 -> ciflow/inductor/162222 2025-09-07T07:36:04.3377922Z * [new tag] ciflow/inductor/162227 -> ciflow/inductor/162227 2025-09-07T07:36:04.3378751Z * [new tag] ciflow/inductor/162238 -> ciflow/inductor/162238 2025-09-07T07:36:04.3379600Z * [new tag] ciflow/inductor/162239 -> ciflow/inductor/162239 2025-09-07T07:36:04.3380429Z * [new tag] ciflow/inductor/162240 -> ciflow/inductor/162240 2025-09-07T07:36:04.3381275Z * [new tag] ciflow/inductor/162244 -> ciflow/inductor/162244 2025-09-07T07:36:04.3382104Z * [new tag] ciflow/inductor/162245 -> ciflow/inductor/162245 2025-09-07T07:36:04.3382943Z * [new tag] ciflow/inductor/162262 -> ciflow/inductor/162262 2025-09-07T07:36:04.3383778Z * [new tag] ciflow/inductor/162275 -> ciflow/inductor/162275 2025-09-07T07:36:04.3384617Z * [new tag] ciflow/inductor/162278 -> ciflow/inductor/162278 2025-09-07T07:36:04.3385456Z * [new tag] ciflow/inductor/162284 -> ciflow/inductor/162284 2025-09-07T07:36:04.3386280Z * [new tag] ciflow/inductor/162286 -> ciflow/inductor/162286 2025-09-07T07:36:04.3387099Z * [new tag] ciflow/inductor/162288 -> ciflow/inductor/162288 2025-09-07T07:36:04.3388077Z * [new tag] ciflow/inductor/162293 -> ciflow/inductor/162293 2025-09-07T07:36:04.3388746Z * [new tag] ciflow/inductor/162294 -> ciflow/inductor/162294 2025-09-07T07:36:04.3389573Z * [new tag] ciflow/inductor/162295 -> ciflow/inductor/162295 2025-09-07T07:36:04.3390405Z * [new tag] ciflow/inductor/162296 -> ciflow/inductor/162296 2025-09-07T07:36:04.3391224Z * [new tag] ciflow/inductor/162298 -> ciflow/inductor/162298 2025-09-07T07:36:04.3392287Z * [new tag] ciflow/inductor/162307 -> ciflow/inductor/162307 2025-09-07T07:36:04.3393143Z * [new tag] ciflow/inductor/162309 -> ciflow/inductor/162309 2025-09-07T07:36:04.3393975Z * [new tag] ciflow/inductor/162311 -> ciflow/inductor/162311 2025-09-07T07:36:04.3394924Z * [new tag] ciflow/inductor/162312 -> ciflow/inductor/162312 2025-09-07T07:36:04.3395681Z * [new tag] ciflow/inductor/162315 -> ciflow/inductor/162315 2025-09-07T07:36:04.3396500Z * [new tag] ciflow/inductor/162316 -> ciflow/inductor/162316 2025-09-07T07:36:04.3397342Z * [new tag] ciflow/inductor/162318 -> ciflow/inductor/162318 2025-09-07T07:36:04.3398613Z * [new tag] ciflow/inductor/162323 -> ciflow/inductor/162323 2025-09-07T07:36:04.3399381Z * [new tag] ciflow/inductor/162341 -> ciflow/inductor/162341 2025-09-07T07:36:04.3400180Z * [new tag] ciflow/inductor/162345 -> ciflow/inductor/162345 2025-09-07T07:36:04.3401341Z * [new tag] ciflow/inductor/3b9a386 -> ciflow/inductor/3b9a386 2025-09-07T07:36:04.3402338Z * [new tag] ciflow/inductor/3d4b92b -> ciflow/inductor/3d4b92b 2025-09-07T07:36:04.3403367Z * [new tag] ciflow/inductor/d224ac7 -> ciflow/inductor/d224ac7 2025-09-07T07:36:04.3404512Z * [new tag] ciflow/linux-aarch64/157994 -> ciflow/linux-aarch64/157994 2025-09-07T07:36:04.3405254Z * [new tag] ciflow/linux-aarch64/159737 -> ciflow/linux-aarch64/159737 2025-09-07T07:36:04.3406022Z * [new tag] ciflow/linux-aarch64/160078 -> ciflow/linux-aarch64/160078 2025-09-07T07:36:04.3407077Z * [new tag] ciflow/mps/157553 -> ciflow/mps/157553 2025-09-07T07:36:04.3407963Z * [new tag] ciflow/mps/157635 -> ciflow/mps/157635 2025-09-07T07:36:04.3408722Z * [new tag] ciflow/mps/161988 -> ciflow/mps/161988 2025-09-07T07:36:04.3409474Z * [new tag] ciflow/mps/162108 -> ciflow/mps/162108 2025-09-07T07:36:04.3410268Z * [new tag] ciflow/mps/162153 -> ciflow/mps/162153 2025-09-07T07:36:04.3411043Z * [new tag] ciflow/mps/162281 -> ciflow/mps/162281 2025-09-07T07:36:04.3412143Z * [new tag] ciflow/nightly/156049 -> ciflow/nightly/156049 2025-09-07T07:36:04.3412826Z * [new tag] ciflow/nightly/158104 -> ciflow/nightly/158104 2025-09-07T07:36:04.3413922Z * [new tag] ciflow/op-benchmark/157994 -> ciflow/op-benchmark/157994 2025-09-07T07:36:04.3415110Z * [new tag] ciflow/periodic-rocm-mi300/161529 -> ciflow/periodic-rocm-mi300/161529 2025-09-07T07:36:04.3415835Z * [new tag] ciflow/periodic-rocm-mi300/161715 -> ciflow/periodic-rocm-mi300/161715 2025-09-07T07:36:04.3417112Z * [new tag] ciflow/periodic/054a2fd -> ciflow/periodic/054a2fd 2025-09-07T07:36:04.3417917Z * [new tag] ciflow/periodic/156703 -> ciflow/periodic/156703 2025-09-07T07:36:04.3418684Z * [new tag] ciflow/periodic/161715 -> ciflow/periodic/161715 2025-09-07T07:36:04.3419466Z * [new tag] ciflow/periodic/162021 -> ciflow/periodic/162021 2025-09-07T07:36:04.3420213Z * [new tag] ciflow/periodic/162323 -> ciflow/periodic/162323 2025-09-07T07:36:04.3421265Z * [new tag] ciflow/periodic/2a6d37d -> ciflow/periodic/2a6d37d 2025-09-07T07:36:04.3422075Z * [new tag] ciflow/periodic/317eeb8 -> ciflow/periodic/317eeb8 2025-09-07T07:36:04.3422976Z * [new tag] ciflow/periodic/3c32 -> ciflow/periodic/3c32 2025-09-07T07:36:04.3424018Z * [new tag] ciflow/periodic/3e98831 -> ciflow/periodic/3e98831 2025-09-07T07:36:04.3425068Z * [new tag] ciflow/periodic/94512-point -> ciflow/periodic/94512-point 2025-09-07T07:36:04.3426218Z * [new tag] ciflow/periodic/csl/test87519 -> ciflow/periodic/csl/test87519 2025-09-07T07:36:04.3427496Z * [new tag] ciflow/periodic/csltest88275 -> ciflow/periodic/csltest88275 2025-09-07T07:36:04.3428559Z * [new tag] ciflow/periodic/csltest88761 -> ciflow/periodic/csltest88761 2025-09-07T07:36:04.3429666Z * [new tag] ciflow/periodic/release_1.12 -> ciflow/periodic/release_1.12 2025-09-07T07:36:04.3430766Z * [new tag] ciflow/periodic/release_1.12.0 -> ciflow/periodic/release_1.12.0 2025-09-07T07:36:04.3431863Z * [new tag] ciflow/periodic/sha-ec5b83 -> ciflow/periodic/sha-ec5b83 2025-09-07T07:36:04.3432877Z * [new tag] ciflow/rocm-mi300/154170 -> ciflow/rocm-mi300/154170 2025-09-07T07:36:04.3433708Z * [new tag] ciflow/rocm-mi300/158747 -> ciflow/rocm-mi300/158747 2025-09-07T07:36:04.3434474Z * [new tag] ciflow/rocm-mi300/159146 -> ciflow/rocm-mi300/159146 2025-09-07T07:36:04.3435267Z * [new tag] ciflow/rocm-mi300/159158 -> ciflow/rocm-mi300/159158 2025-09-07T07:36:04.3436015Z * [new tag] ciflow/rocm-mi300/161715 -> ciflow/rocm-mi300/161715 2025-09-07T07:36:04.3436813Z * [new tag] ciflow/rocm-mi300/161957 -> ciflow/rocm-mi300/161957 2025-09-07T07:36:04.3437773Z * [new tag] ciflow/rocm-mi300/162053 -> ciflow/rocm-mi300/162053 2025-09-07T07:36:04.3438962Z * [new tag] ciflow/rocm-mi300/162056 -> ciflow/rocm-mi300/162056 2025-09-07T07:36:04.3439805Z * [new tag] ciflow/rocm-mi300/162112 -> ciflow/rocm-mi300/162112 2025-09-07T07:36:04.3440578Z * [new tag] ciflow/rocm-mi300/162245 -> ciflow/rocm-mi300/162245 2025-09-07T07:36:04.3441422Z * [new tag] ciflow/rocm-mi300/162278 -> ciflow/rocm-mi300/162278 2025-09-07T07:36:04.3442465Z * [new tag] ciflow/rocm-mi300/162288 -> ciflow/rocm-mi300/162288 2025-09-07T07:36:04.3443512Z * [new tag] ciflow/rocm-mi355/162053 -> ciflow/rocm-mi355/162053 2025-09-07T07:36:04.3444316Z * [new tag] ciflow/rocm-mi355/162056 -> ciflow/rocm-mi355/162056 2025-09-07T07:36:04.3445379Z * [new tag] ciflow/rocm/148492 -> ciflow/rocm/148492 2025-09-07T07:36:04.3446095Z * [new tag] ciflow/rocm/154170 -> ciflow/rocm/154170 2025-09-07T07:36:04.3447107Z * [new tag] ciflow/rocm/156491 -> ciflow/rocm/156491 2025-09-07T07:36:04.3447830Z * [new tag] ciflow/rocm/156592 -> ciflow/rocm/156592 2025-09-07T07:36:04.3448592Z * [new tag] ciflow/rocm/158747 -> ciflow/rocm/158747 2025-09-07T07:36:04.3449342Z * [new tag] ciflow/rocm/159146 -> ciflow/rocm/159146 2025-09-07T07:36:04.3450366Z * [new tag] ciflow/rocm/159158 -> ciflow/rocm/159158 2025-09-07T07:36:04.3451065Z * [new tag] ciflow/rocm/161715 -> ciflow/rocm/161715 2025-09-07T07:36:04.3452033Z * [new tag] ciflow/rocm/161972 -> ciflow/rocm/161972 2025-09-07T07:36:04.3452735Z * [new tag] ciflow/rocm/162052 -> ciflow/rocm/162052 2025-09-07T07:36:04.3453536Z * [new tag] ciflow/rocm/162053 -> ciflow/rocm/162053 2025-09-07T07:36:04.3454697Z * [new tag] ciflow/rocm/162056 -> ciflow/rocm/162056 2025-09-07T07:36:04.3455754Z * [new tag] ciflow/rocm/162112 -> ciflow/rocm/162112 2025-09-07T07:36:04.3456663Z * [new tag] ciflow/rocm/162278 -> ciflow/rocm/162278 2025-09-07T07:36:04.3458075Z * [new tag] ciflow/rocm/162288 -> ciflow/rocm/162288 2025-09-07T07:36:04.3458876Z * [new tag] ciflow/rocm/162305 -> ciflow/rocm/162305 2025-09-07T07:36:04.3460090Z * [new tag] ciflow/slow/01c7106 -> ciflow/slow/01c7106 2025-09-07T07:36:04.3460905Z * [new tag] ciflow/slow/0577043 -> ciflow/slow/0577043 2025-09-07T07:36:04.3462277Z * [new tag] ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym -> ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym 2025-09-07T07:36:04.3462915Z * [new tag] ciflow/slow/0e81104 -> ciflow/slow/0e81104 2025-09-07T07:36:04.3463713Z * [new tag] ciflow/slow/161395 -> ciflow/slow/161395 2025-09-07T07:36:04.3464696Z * [new tag] ciflow/slow/1732077 -> ciflow/slow/1732077 2025-09-07T07:36:04.3465707Z * [new tag] ciflow/slow/187eb7c -> ciflow/slow/187eb7c 2025-09-07T07:36:04.3466458Z * [new tag] ciflow/slow/1faef89 -> ciflow/slow/1faef89 2025-09-07T07:36:04.3467695Z * [new tag] ciflow/slow/3920ec1 -> ciflow/slow/3920ec1 2025-09-07T07:36:04.3468812Z * [new tag] ciflow/slow/3b7c6b2 -> ciflow/slow/3b7c6b2 2025-09-07T07:36:04.3469808Z * [new tag] ciflow/slow/59a3759 -> ciflow/slow/59a3759 2025-09-07T07:36:04.3470767Z * [new tag] ciflow/slow/70ef0bb -> ciflow/slow/70ef0bb 2025-09-07T07:36:04.3471705Z * [new tag] ciflow/slow/788ff06 -> ciflow/slow/788ff06 2025-09-07T07:36:04.3473041Z * [new tag] ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym -> ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym 2025-09-07T07:36:04.3473721Z * [new tag] ciflow/slow/9d85864 -> ciflow/slow/9d85864 2025-09-07T07:36:04.3474777Z * [new tag] ciflow/slow/9ffad5b -> ciflow/slow/9ffad5b 2025-09-07T07:36:04.3475802Z * [new tag] ciflow/slow/a206e8b -> ciflow/slow/a206e8b 2025-09-07T07:36:04.3477063Z * [new tag] ciflow/slow/a837609 -> ciflow/slow/a837609 2025-09-07T07:36:04.3478059Z * [new tag] ciflow/slow/af841f3 -> ciflow/slow/af841f3 2025-09-07T07:36:04.3479290Z * [new tag] ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym -> ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym 2025-09-07T07:36:04.3480095Z * [new tag] ciflow/triton_binaries/162329 -> ciflow/triton_binaries/162329 2025-09-07T07:36:04.3481111Z * [new tag] ciflow/trunk/113258 -> ciflow/trunk/113258 2025-09-07T07:36:04.3481824Z * [new tag] ciflow/trunk/137400 -> ciflow/trunk/137400 2025-09-07T07:36:04.3482589Z * [new tag] ciflow/trunk/148180 -> ciflow/trunk/148180 2025-09-07T07:36:04.3483376Z * [new tag] ciflow/trunk/148328 -> ciflow/trunk/148328 2025-09-07T07:36:04.3484133Z * [new tag] ciflow/trunk/148492 -> ciflow/trunk/148492 2025-09-07T07:36:04.3485335Z * [new tag] ciflow/trunk/148919 -> ciflow/trunk/148919 2025-09-07T07:36:04.3486045Z * [new tag] ciflow/trunk/152624 -> ciflow/trunk/152624 2025-09-07T07:36:04.3486826Z * [new tag] ciflow/trunk/154170 -> ciflow/trunk/154170 2025-09-07T07:36:04.3487573Z * [new tag] ciflow/trunk/154694 -> ciflow/trunk/154694 2025-09-07T07:36:04.3488339Z * [new tag] ciflow/trunk/156049 -> ciflow/trunk/156049 2025-09-07T07:36:04.3489102Z * [new tag] ciflow/trunk/156703 -> ciflow/trunk/156703 2025-09-07T07:36:04.3490202Z * [new tag] ciflow/trunk/156711 -> ciflow/trunk/156711 2025-09-07T07:36:04.3491245Z * [new tag] ciflow/trunk/157432 -> ciflow/trunk/157432 2025-09-07T07:36:04.3492254Z * [new tag] ciflow/trunk/157685 -> ciflow/trunk/157685 2025-09-07T07:36:04.3493013Z * [new tag] ciflow/trunk/157689 -> ciflow/trunk/157689 2025-09-07T07:36:04.3493857Z * [new tag] ciflow/trunk/157699 -> ciflow/trunk/157699 2025-09-07T07:36:04.3494679Z * [new tag] ciflow/trunk/157813 -> ciflow/trunk/157813 2025-09-07T07:36:04.3495490Z * [new tag] ciflow/trunk/157994 -> ciflow/trunk/157994 2025-09-07T07:36:04.3496347Z * [new tag] ciflow/trunk/158091 -> ciflow/trunk/158091 2025-09-07T07:36:04.3497175Z * [new tag] ciflow/trunk/158104 -> ciflow/trunk/158104 2025-09-07T07:36:04.3498500Z * [new tag] ciflow/trunk/158404 -> ciflow/trunk/158404 2025-09-07T07:36:04.3499316Z * [new tag] ciflow/trunk/158647 -> ciflow/trunk/158647 2025-09-07T07:36:04.3500458Z * [new tag] ciflow/trunk/158846 -> ciflow/trunk/158846 2025-09-07T07:36:04.3501217Z * [new tag] ciflow/trunk/159158 -> ciflow/trunk/159158 2025-09-07T07:36:04.3502202Z * [new tag] ciflow/trunk/159682 -> ciflow/trunk/159682 2025-09-07T07:36:04.3502959Z * [new tag] ciflow/trunk/159835 -> ciflow/trunk/159835 2025-09-07T07:36:04.3504064Z * [new tag] ciflow/trunk/160161 -> ciflow/trunk/160161 2025-09-07T07:36:04.3504837Z * [new tag] ciflow/trunk/160236 -> ciflow/trunk/160236 2025-09-07T07:36:04.3505655Z * [new tag] ciflow/trunk/160329 -> ciflow/trunk/160329 2025-09-07T07:36:04.3506469Z * [new tag] ciflow/trunk/160480 -> ciflow/trunk/160480 2025-09-07T07:36:04.3507296Z * [new tag] ciflow/trunk/160532 -> ciflow/trunk/160532 2025-09-07T07:36:04.3508088Z * [new tag] ciflow/trunk/160836 -> ciflow/trunk/160836 2025-09-07T07:36:04.3509034Z * [new tag] ciflow/trunk/160843 -> ciflow/trunk/160843 2025-09-07T07:36:04.3509749Z * [new tag] ciflow/trunk/160869 -> ciflow/trunk/160869 2025-09-07T07:36:04.3510796Z * [new tag] ciflow/trunk/160940 -> ciflow/trunk/160940 2025-09-07T07:36:04.3511572Z * [new tag] ciflow/trunk/160943 -> ciflow/trunk/160943 2025-09-07T07:36:04.3512652Z * [new tag] ciflow/trunk/160953 -> ciflow/trunk/160953 2025-09-07T07:36:04.3513635Z * [new tag] ciflow/trunk/161035 -> ciflow/trunk/161035 2025-09-07T07:36:04.3514459Z * [new tag] ciflow/trunk/161178 -> ciflow/trunk/161178 2025-09-07T07:36:04.3515274Z * [new tag] ciflow/trunk/161349 -> ciflow/trunk/161349 2025-09-07T07:36:04.3516105Z * [new tag] ciflow/trunk/161350 -> ciflow/trunk/161350 2025-09-07T07:36:04.3516924Z * [new tag] ciflow/trunk/161351 -> ciflow/trunk/161351 2025-09-07T07:36:04.3517732Z * [new tag] ciflow/trunk/161395 -> ciflow/trunk/161395 2025-09-07T07:36:04.3518534Z * [new tag] ciflow/trunk/161405 -> ciflow/trunk/161405 2025-09-07T07:36:04.3519348Z * [new tag] ciflow/trunk/161406 -> ciflow/trunk/161406 2025-09-07T07:36:04.3520167Z * [new tag] ciflow/trunk/161410 -> ciflow/trunk/161410 2025-09-07T07:36:04.3521001Z * [new tag] ciflow/trunk/161468 -> ciflow/trunk/161468 2025-09-07T07:36:04.3521824Z * [new tag] ciflow/trunk/161499 -> ciflow/trunk/161499 2025-09-07T07:36:04.3522957Z * [new tag] ciflow/trunk/161527 -> ciflow/trunk/161527 2025-09-07T07:36:04.3523786Z * [new tag] ciflow/trunk/161534 -> ciflow/trunk/161534 2025-09-07T07:36:04.3524563Z * [new tag] ciflow/trunk/161591 -> ciflow/trunk/161591 2025-09-07T07:36:04.3525368Z * [new tag] ciflow/trunk/161595 -> ciflow/trunk/161595 2025-09-07T07:36:04.3526195Z * [new tag] ciflow/trunk/161596 -> ciflow/trunk/161596 2025-09-07T07:36:04.3527017Z * [new tag] ciflow/trunk/161633 -> ciflow/trunk/161633 2025-09-07T07:36:04.3527965Z * [new tag] ciflow/trunk/161634 -> ciflow/trunk/161634 2025-09-07T07:36:04.3528925Z * [new tag] ciflow/trunk/161635 -> ciflow/trunk/161635 2025-09-07T07:36:04.3529703Z * [new tag] ciflow/trunk/161667 -> ciflow/trunk/161667 2025-09-07T07:36:04.3530517Z * [new tag] ciflow/trunk/161670 -> ciflow/trunk/161670 2025-09-07T07:36:04.3531346Z * [new tag] ciflow/trunk/161692 -> ciflow/trunk/161692 2025-09-07T07:36:04.3532176Z * [new tag] ciflow/trunk/161693 -> ciflow/trunk/161693 2025-09-07T07:36:04.3533110Z * [new tag] ciflow/trunk/161695 -> ciflow/trunk/161695 2025-09-07T07:36:04.3533875Z * [new tag] ciflow/trunk/161730 -> ciflow/trunk/161730 2025-09-07T07:36:04.3534687Z * [new tag] ciflow/trunk/161744 -> ciflow/trunk/161744 2025-09-07T07:36:04.3535497Z * [new tag] ciflow/trunk/161749 -> ciflow/trunk/161749 2025-09-07T07:36:04.3536326Z * [new tag] ciflow/trunk/161881 -> ciflow/trunk/161881 2025-09-07T07:36:04.3537137Z * [new tag] ciflow/trunk/161924 -> ciflow/trunk/161924 2025-09-07T07:36:04.3538228Z * [new tag] ciflow/trunk/161926 -> ciflow/trunk/161926 2025-09-07T07:36:04.3539010Z * [new tag] ciflow/trunk/161936 -> ciflow/trunk/161936 2025-09-07T07:36:04.3539837Z * [new tag] ciflow/trunk/161952 -> ciflow/trunk/161952 2025-09-07T07:36:04.3540770Z * [new tag] ciflow/trunk/161955 -> ciflow/trunk/161955 2025-09-07T07:36:04.3541487Z * [new tag] ciflow/trunk/161957 -> ciflow/trunk/161957 2025-09-07T07:36:04.3542298Z * [new tag] ciflow/trunk/161959 -> ciflow/trunk/161959 2025-09-07T07:36:04.3543109Z * [new tag] ciflow/trunk/161977 -> ciflow/trunk/161977 2025-09-07T07:36:04.3543920Z * [new tag] ciflow/trunk/161988 -> ciflow/trunk/161988 2025-09-07T07:36:04.3544756Z * [new tag] ciflow/trunk/161994 -> ciflow/trunk/161994 2025-09-07T07:36:04.3545761Z * [new tag] ciflow/trunk/162007 -> ciflow/trunk/162007 2025-09-07T07:36:04.3546519Z * [new tag] ciflow/trunk/162013 -> ciflow/trunk/162013 2025-09-07T07:36:04.3547342Z * [new tag] ciflow/trunk/162017 -> ciflow/trunk/162017 2025-09-07T07:36:04.3548176Z * [new tag] ciflow/trunk/162021 -> ciflow/trunk/162021 2025-09-07T07:36:04.3549008Z * [new tag] ciflow/trunk/162022 -> ciflow/trunk/162022 2025-09-07T07:36:04.3549850Z * [new tag] ciflow/trunk/162040 -> ciflow/trunk/162040 2025-09-07T07:36:04.3550677Z * [new tag] ciflow/trunk/162041 -> ciflow/trunk/162041 2025-09-07T07:36:04.3551679Z * [new tag] ciflow/trunk/162062 -> ciflow/trunk/162062 2025-09-07T07:36:04.3552457Z * [new tag] ciflow/trunk/162066 -> ciflow/trunk/162066 2025-09-07T07:36:04.3553701Z * [new tag] ciflow/trunk/162089 -> ciflow/trunk/162089 2025-09-07T07:36:04.3554486Z * [new tag] ciflow/trunk/162099 -> ciflow/trunk/162099 2025-09-07T07:36:04.3555452Z * [new tag] ciflow/trunk/162104 -> ciflow/trunk/162104 2025-09-07T07:36:04.3556231Z * [new tag] ciflow/trunk/162106 -> ciflow/trunk/162106 2025-09-07T07:36:04.3557056Z * [new tag] ciflow/trunk/162112 -> ciflow/trunk/162112 2025-09-07T07:36:04.3557889Z * [new tag] ciflow/trunk/162119 -> ciflow/trunk/162119 2025-09-07T07:36:04.3558734Z * [new tag] ciflow/trunk/162142 -> ciflow/trunk/162142 2025-09-07T07:36:04.3559552Z * [new tag] ciflow/trunk/162169 -> ciflow/trunk/162169 2025-09-07T07:36:04.3560363Z * [new tag] ciflow/trunk/162183 -> ciflow/trunk/162183 2025-09-07T07:36:04.3561174Z * [new tag] ciflow/trunk/162190 -> ciflow/trunk/162190 2025-09-07T07:36:04.3561989Z * [new tag] ciflow/trunk/162194 -> ciflow/trunk/162194 2025-09-07T07:36:04.3562808Z * [new tag] ciflow/trunk/162200 -> ciflow/trunk/162200 2025-09-07T07:36:04.3563637Z * [new tag] ciflow/trunk/162206 -> ciflow/trunk/162206 2025-09-07T07:36:04.3564478Z * [new tag] ciflow/trunk/162208 -> ciflow/trunk/162208 2025-09-07T07:36:04.3565482Z * [new tag] ciflow/trunk/162222 -> ciflow/trunk/162222 2025-09-07T07:36:04.3566249Z * [new tag] ciflow/trunk/162238 -> ciflow/trunk/162238 2025-09-07T07:36:04.3567064Z * [new tag] ciflow/trunk/162244 -> ciflow/trunk/162244 2025-09-07T07:36:04.3568211Z * [new tag] ciflow/trunk/162267 -> ciflow/trunk/162267 2025-09-07T07:36:04.3569291Z * [new tag] ciflow/trunk/162269 -> ciflow/trunk/162269 2025-09-07T07:36:04.3569932Z * [new tag] ciflow/trunk/162278 -> ciflow/trunk/162278 2025-09-07T07:36:04.3570990Z * [new tag] ciflow/trunk/162286 -> ciflow/trunk/162286 2025-09-07T07:36:04.3571752Z * [new tag] ciflow/trunk/162288 -> ciflow/trunk/162288 2025-09-07T07:36:04.3572598Z * [new tag] ciflow/trunk/162293 -> ciflow/trunk/162293 2025-09-07T07:36:04.3573491Z * [new tag] ciflow/trunk/162310 -> ciflow/trunk/162310 2025-09-07T07:36:04.3574243Z * [new tag] ciflow/trunk/162311 -> ciflow/trunk/162311 2025-09-07T07:36:04.3575065Z * [new tag] ciflow/trunk/162315 -> ciflow/trunk/162315 2025-09-07T07:36:04.3575882Z * [new tag] ciflow/trunk/162325 -> ciflow/trunk/162325 2025-09-07T07:36:04.3576904Z * [new tag] ciflow/trunk/162328 -> ciflow/trunk/162328 2025-09-07T07:36:04.3577732Z * [new tag] ciflow/trunk/162329 -> ciflow/trunk/162329 2025-09-07T07:36:04.3579075Z * [new tag] ciflow/unstable/123 -> ciflow/unstable/123 2025-09-07T07:36:04.3580111Z * [new tag] ciflow/vllm/162292 -> ciflow/vllm/162292 2025-09-07T07:36:04.3593375Z * [new tag] ciflow/win-arm64/156049 -> ciflow/win-arm64/156049 2025-09-07T07:36:04.3593710Z * [new tag] ciflow/win-arm64/158104 -> ciflow/win-arm64/158104 2025-09-07T07:36:04.3594017Z * [new tag] ciflow/xpu/157699 -> ciflow/xpu/157699 2025-09-07T07:36:04.3594294Z * [new tag] ciflow/xpu/157994 -> ciflow/xpu/157994 2025-09-07T07:36:04.3594543Z * [new tag] ciflow/xpu/159459 -> ciflow/xpu/159459 2025-09-07T07:36:04.3594799Z * [new tag] ciflow/xpu/159718 -> ciflow/xpu/159718 2025-09-07T07:36:04.3595053Z * [new tag] ciflow/xpu/159944 -> ciflow/xpu/159944 2025-09-07T07:36:04.3595308Z * [new tag] ciflow/xpu/160867 -> ciflow/xpu/160867 2025-09-07T07:36:04.3595555Z * [new tag] ciflow/xpu/160938 -> ciflow/xpu/160938 2025-09-07T07:36:04.3595806Z * [new tag] ciflow/xpu/160940 -> ciflow/xpu/160940 2025-09-07T07:36:04.3596171Z * [new tag] ciflow/xpu/160953 -> ciflow/xpu/160953 2025-09-07T07:36:04.3596429Z * [new tag] ciflow/xpu/161045 -> ciflow/xpu/161045 2025-09-07T07:36:04.3596683Z * [new tag] ciflow/xpu/161058 -> ciflow/xpu/161058 2025-09-07T07:36:04.3596931Z * [new tag] ciflow/xpu/161246 -> ciflow/xpu/161246 2025-09-07T07:36:04.3597182Z * [new tag] ciflow/xpu/161397 -> ciflow/xpu/161397 2025-09-07T07:36:04.3597443Z * [new tag] ciflow/xpu/161485 -> ciflow/xpu/161485 2025-09-07T07:36:04.3597707Z * [new tag] ciflow/xpu/161988 -> ciflow/xpu/161988 2025-09-07T07:36:04.3597965Z * [new tag] ciflow/xpu/162062 -> ciflow/xpu/162062 2025-09-07T07:36:04.3598216Z * [new tag] cslpull75 -> cslpull75 2025-09-07T07:36:04.3598626Z * [new tag] cslpull76 -> cslpull76 2025-09-07T07:36:04.3599423Z * [new tag] cslpull77 -> cslpull77 2025-09-07T07:36:04.3600255Z * [new tag] cslpull78 -> cslpull78 2025-09-07T07:36:04.3601409Z * [new tag] cslpull79 -> cslpull79 2025-09-07T07:36:04.3602566Z * [new tag] cslpull80 -> cslpull80 2025-09-07T07:36:04.3606305Z * [new tag] cslpull81 -> cslpull81 2025-09-07T07:36:04.3607311Z * [new tag] cslpull82 -> cslpull82 2025-09-07T07:36:04.3608149Z * [new tag] cslpull83 -> cslpull83 2025-09-07T07:36:04.3609126Z * [new tag] cslpull84 -> cslpull84 2025-09-07T07:36:04.3610133Z * [new tag] cslpull85 -> cslpull85 2025-09-07T07:36:04.3610896Z * [new tag] cslpull86 -> cslpull86 2025-09-07T07:36:04.3611833Z * [new tag] cslpull87 -> cslpull87 2025-09-07T07:36:04.3612975Z * [new tag] cslpull88 -> cslpull88 2025-09-07T07:36:04.3613784Z * [new tag] cslpull89 -> cslpull89 2025-09-07T07:36:04.3614574Z * [new tag] cslpull90 -> cslpull90 2025-09-07T07:36:04.3615788Z * [new tag] cslpull91 -> cslpull91 2025-09-07T07:36:04.3616604Z * [new tag] cslpull92 -> cslpull92 2025-09-07T07:36:04.3617601Z * [new tag] flight_5 -> flight_5 2025-09-07T07:36:04.3618684Z * [new tag] flight_5.1 -> flight_5.1 2025-09-07T07:36:04.3619656Z * [new tag] flight_5.2 -> flight_5.2 2025-09-07T07:36:04.3620840Z * [new tag] flight_5.3 -> flight_5.3 2025-09-07T07:36:04.3621570Z * [new tag] forpull1 -> forpull1 2025-09-07T07:36:04.3622774Z * [new tag] malfet/tag-2ef5611 -> malfet/tag-2ef5611 2025-09-07T07:36:04.3623610Z * [new tag] malfet/tag-317b1a0 -> malfet/tag-317b1a0 2025-09-07T07:36:04.3624575Z * [new tag] malfet/tag-ec6f767 -> malfet/tag-ec6f767 2025-09-07T07:36:04.3625557Z * [new tag] nightly-binary -> nightly-binary 2025-09-07T07:36:04.3626268Z * [new tag] sqzhang_flight4_plus -> sqzhang_flight4_plus 2025-09-07T07:36:04.3627305Z * [new tag] sqzhang_flight_3 -> sqzhang_flight_3 2025-09-07T07:36:04.3628619Z * [new tag] trunk/00636e0171e7e733628c408084805442270cf608 -> trunk/00636e0171e7e733628c408084805442270cf608 2025-09-07T07:36:04.3629419Z * [new tag] trunk/019fed39aa6b2dd8c69347378d53423e5efae8d4 -> trunk/019fed39aa6b2dd8c69347378d53423e5efae8d4 2025-09-07T07:36:04.3630612Z * [new tag] trunk/01ab325cc2e0dc221af4d710974e1b9175066544 -> trunk/01ab325cc2e0dc221af4d710974e1b9175066544 2025-09-07T07:36:04.3631726Z * [new tag] trunk/01edcd4df8bf0c7b4cc2d3ec868bd2059eeea83b -> trunk/01edcd4df8bf0c7b4cc2d3ec868bd2059eeea83b 2025-09-07T07:36:04.3632725Z * [new tag] trunk/040d00af048967dde7938d358d7f5988cbd18388 -> trunk/040d00af048967dde7938d358d7f5988cbd18388 2025-09-07T07:36:04.3633716Z * [new tag] trunk/0447f2d99b4351b2ff129dce6eebb371024f73e5 -> trunk/0447f2d99b4351b2ff129dce6eebb371024f73e5 2025-09-07T07:36:04.3634714Z * [new tag] trunk/047603d35bdc70046216384838d6340feab79bf4 -> trunk/047603d35bdc70046216384838d6340feab79bf4 2025-09-07T07:36:04.3635665Z * [new tag] trunk/06da7c0730b3764f178ec3a90dedf4ffa4202d81 -> trunk/06da7c0730b3764f178ec3a90dedf4ffa4202d81 2025-09-07T07:36:04.3636738Z * [new tag] trunk/081cab045472ce045634548cc6c14a4870641e23 -> trunk/081cab045472ce045634548cc6c14a4870641e23 2025-09-07T07:36:04.3637682Z * [new tag] trunk/09587daf8c9f21f5340f73921ce5f23d1a4a4572 -> trunk/09587daf8c9f21f5340f73921ce5f23d1a4a4572 2025-09-07T07:36:04.3638596Z * [new tag] trunk/09be1890d72cc34fc946965dc4a27736bf0ca8c6 -> trunk/09be1890d72cc34fc946965dc4a27736bf0ca8c6 2025-09-07T07:36:04.3639544Z * [new tag] trunk/09d2f1b6315d6d416fbf452793d65795863ebc66 -> trunk/09d2f1b6315d6d416fbf452793d65795863ebc66 2025-09-07T07:36:04.3640435Z * [new tag] trunk/0af70e2353e1dcda83175fd4834ecb7b63e009e0 -> trunk/0af70e2353e1dcda83175fd4834ecb7b63e009e0 2025-09-07T07:36:04.3641951Z * [new tag] trunk/0c0e056a9e20c17271a6144dd32c0c7e3ba26736 -> trunk/0c0e056a9e20c17271a6144dd32c0c7e3ba26736 2025-09-07T07:36:04.3642921Z * [new tag] trunk/0cd6c56bdfa9178ff61be82ce3b178926ddb64a9 -> trunk/0cd6c56bdfa9178ff61be82ce3b178926ddb64a9 2025-09-07T07:36:04.3643840Z * [new tag] trunk/0d421ace32c1605ee8e452ee1eeb03bd243dd96c -> trunk/0d421ace32c1605ee8e452ee1eeb03bd243dd96c 2025-09-07T07:36:04.3644986Z * [new tag] trunk/0d71a9dd5b4b6d1dde58d91c9b71d96bc6a6a171 -> trunk/0d71a9dd5b4b6d1dde58d91c9b71d96bc6a6a171 2025-09-07T07:36:04.3645835Z * [new tag] trunk/0d84ff3b78f55492d3d4708458c92d776274939e -> trunk/0d84ff3b78f55492d3d4708458c92d776274939e 2025-09-07T07:36:04.3646758Z * [new tag] trunk/0f45aaf4414048b17d720d0915ce221a8de8ec63 -> trunk/0f45aaf4414048b17d720d0915ce221a8de8ec63 2025-09-07T07:36:04.3647692Z * [new tag] trunk/0ff8eabf1387de5acd6712a03bda61f1a3dfa27f -> trunk/0ff8eabf1387de5acd6712a03bda61f1a3dfa27f 2025-09-07T07:36:04.3648637Z * [new tag] trunk/104f2680e03d13a4765ca69f905d8f16fc0c822f -> trunk/104f2680e03d13a4765ca69f905d8f16fc0c822f 2025-09-07T07:36:04.3649600Z * [new tag] trunk/12814701555d3e41dfcdf8f9273af5821e322df0 -> trunk/12814701555d3e41dfcdf8f9273af5821e322df0 2025-09-07T07:36:04.3650574Z * [new tag] trunk/13b65196db422bdb394cb482e208c61ed448898c -> trunk/13b65196db422bdb394cb482e208c61ed448898c 2025-09-07T07:36:04.3651482Z * [new tag] trunk/13d66e2a66eceed14b8a8f5a971087df4f688a46 -> trunk/13d66e2a66eceed14b8a8f5a971087df4f688a46 2025-09-07T07:36:04.3652398Z * [new tag] trunk/145a3a7bda15e3963a33eb1b54bba5d4a270b225 -> trunk/145a3a7bda15e3963a33eb1b54bba5d4a270b225 2025-09-07T07:36:04.3653349Z * [new tag] trunk/146371483318e17929daefd37c8e459d9d6d47bb -> trunk/146371483318e17929daefd37c8e459d9d6d47bb 2025-09-07T07:36:04.3654334Z * [new tag] trunk/15c77a8cfd341e74fd124b077492ef2bfa51b339 -> trunk/15c77a8cfd341e74fd124b077492ef2bfa51b339 2025-09-07T07:36:04.3655274Z * [new tag] trunk/17fa8eec4a1e32939ab4d364ee6e75487a79b654 -> trunk/17fa8eec4a1e32939ab4d364ee6e75487a79b654 2025-09-07T07:36:04.3656689Z * [new tag] trunk/190c391a28845a14df26abb228d26aa813efb20c -> trunk/190c391a28845a14df26abb228d26aa813efb20c 2025-09-07T07:36:04.3657665Z * [new tag] trunk/1a588ace4667bde1331fbd8ed957157dca5cee68 -> trunk/1a588ace4667bde1331fbd8ed957157dca5cee68 2025-09-07T07:36:04.3658666Z * [new tag] trunk/1aa7476885e8f6e7b0ec3a5b6383aad9d3f343e7 -> trunk/1aa7476885e8f6e7b0ec3a5b6383aad9d3f343e7 2025-09-07T07:36:04.3659474Z * [new tag] trunk/1aeb421c342c9e9607842f4c87cb46e8e816ee53 -> trunk/1aeb421c342c9e9607842f4c87cb46e8e816ee53 2025-09-07T07:36:04.3660412Z * [new tag] trunk/1c1b28d5b6a942fafe23b2f09302d93c25226d4a -> trunk/1c1b28d5b6a942fafe23b2f09302d93c25226d4a 2025-09-07T07:36:04.3661320Z * [new tag] trunk/1ebd70d0c0d562d3be9abdee2a21906584af7d99 -> trunk/1ebd70d0c0d562d3be9abdee2a21906584af7d99 2025-09-07T07:36:04.3662265Z * [new tag] trunk/1ec2c15914da4ef7bd926ed9aebc8671c75fe965 -> trunk/1ec2c15914da4ef7bd926ed9aebc8671c75fe965 2025-09-07T07:36:04.3663203Z * [new tag] trunk/1f51056bd64e73d1aa81321bc3c098575b1bc78a -> trunk/1f51056bd64e73d1aa81321bc3c098575b1bc78a 2025-09-07T07:36:04.3664199Z * [new tag] trunk/1f820de639c75a1562d3fb03f160439f853ae07b -> trunk/1f820de639c75a1562d3fb03f160439f853ae07b 2025-09-07T07:36:04.3665258Z * [new tag] trunk/204697f0e695d82894c5010fbec664c4391f90cc -> trunk/204697f0e695d82894c5010fbec664c4391f90cc 2025-09-07T07:36:04.3666153Z * [new tag] trunk/20629b1619fe636227d01fc85ba221daa7185a05 -> trunk/20629b1619fe636227d01fc85ba221daa7185a05 2025-09-07T07:36:04.3667088Z * [new tag] trunk/20b47acef845e9c4f71da9429a396d293f50ebe7 -> trunk/20b47acef845e9c4f71da9429a396d293f50ebe7 2025-09-07T07:36:04.3668026Z * [new tag] trunk/20bfb2539d7c5250379648eda35f80b8a7d642dd -> trunk/20bfb2539d7c5250379648eda35f80b8a7d642dd 2025-09-07T07:36:04.3669028Z * [new tag] trunk/21fae99c180d17def562797ea0fb154d8fdf88e3 -> trunk/21fae99c180d17def562797ea0fb154d8fdf88e3 2025-09-07T07:36:04.3670235Z * [new tag] trunk/248355faf53f9f7ba2fd0a367d59600c6d991e7f -> trunk/248355faf53f9f7ba2fd0a367d59600c6d991e7f 2025-09-07T07:36:04.3671105Z * [new tag] trunk/25f4aaed9ec26f39c13862323ff8582006473d23 -> trunk/25f4aaed9ec26f39c13862323ff8582006473d23 2025-09-07T07:36:04.3672073Z * [new tag] trunk/261a84a1764412f8e659c956e3f81997ec3de9d5 -> trunk/261a84a1764412f8e659c956e3f81997ec3de9d5 2025-09-07T07:36:04.3673117Z * [new tag] trunk/28f4ab0737937858730f29f5c4e601e109cf9d5f -> trunk/28f4ab0737937858730f29f5c4e601e109cf9d5f 2025-09-07T07:36:04.3674140Z * [new tag] trunk/291cd11f2d5df6f48d348cce0e4e762f274f4dc4 -> trunk/291cd11f2d5df6f48d348cce0e4e762f274f4dc4 2025-09-07T07:36:04.3675080Z * [new tag] trunk/29280864d941e6108ab57f7298f520c0cf9696e9 -> trunk/29280864d941e6108ab57f7298f520c0cf9696e9 2025-09-07T07:36:04.3676118Z * [new tag] trunk/2a45837e98c63cae9d1a2e2133a727b829e549d5 -> trunk/2a45837e98c63cae9d1a2e2133a727b829e549d5 2025-09-07T07:36:04.3677177Z * [new tag] trunk/2a5c0785e2f975697fd7bdf1411de6e03dcaa1ef -> trunk/2a5c0785e2f975697fd7bdf1411de6e03dcaa1ef 2025-09-07T07:36:04.3678154Z * [new tag] trunk/2b8a83901c58a0858ea9e4ce00055f48e6ed164c -> trunk/2b8a83901c58a0858ea9e4ce00055f48e6ed164c 2025-09-07T07:36:04.3679060Z * [new tag] trunk/2ba65472dd54488a86a50326ea990195fc6732d6 -> trunk/2ba65472dd54488a86a50326ea990195fc6732d6 2025-09-07T07:36:04.3680001Z * [new tag] trunk/2c03f0acc53ed13fe8ebfe809129f25996e009a0 -> trunk/2c03f0acc53ed13fe8ebfe809129f25996e009a0 2025-09-07T07:36:04.3680905Z * [new tag] trunk/2dd529df0092799f68ee7afcf52338276906706a -> trunk/2dd529df0092799f68ee7afcf52338276906706a 2025-09-07T07:36:04.3681868Z * [new tag] trunk/2f6b4b1ad3f82bb3bd984f6e65744ea339ffb8b5 -> trunk/2f6b4b1ad3f82bb3bd984f6e65744ea339ffb8b5 2025-09-07T07:36:04.3682778Z * [new tag] trunk/2fa0520a64ed8aa734a56c4d124958f0b5711ca8 -> trunk/2fa0520a64ed8aa734a56c4d124958f0b5711ca8 2025-09-07T07:36:04.3683699Z * [new tag] trunk/302df2ac5dc4222294c09d48804a2dddb8f4bad8 -> trunk/302df2ac5dc4222294c09d48804a2dddb8f4bad8 2025-09-07T07:36:04.3684516Z * [new tag] trunk/33028597bfa2e0178e28c8cce33cb9b3800cac43 -> trunk/33028597bfa2e0178e28c8cce33cb9b3800cac43 2025-09-07T07:36:04.3685517Z * [new tag] trunk/34aa78274d6770086025a967fa63a86830e08176 -> trunk/34aa78274d6770086025a967fa63a86830e08176 2025-09-07T07:36:04.3686442Z * [new tag] trunk/3559c354ce6a14d11fe29fb12fa2747a2f2af449 -> trunk/3559c354ce6a14d11fe29fb12fa2747a2f2af449 2025-09-07T07:36:04.3687804Z * [new tag] trunk/36d207fcaaede0d1e58a5168084c307b32b6fd8b -> trunk/36d207fcaaede0d1e58a5168084c307b32b6fd8b 2025-09-07T07:36:04.3688605Z * [new tag] trunk/377033757ae5ca524ea842f1b0a5f446ed3d8fe0 -> trunk/377033757ae5ca524ea842f1b0a5f446ed3d8fe0 2025-09-07T07:36:04.3689655Z * [new tag] trunk/3771380f83fcac154a7c89ad679311d8c4818287 -> trunk/3771380f83fcac154a7c89ad679311d8c4818287 2025-09-07T07:36:04.3690583Z * [new tag] trunk/3a207816cc569f78863d86c01f2a3d265350e39f -> trunk/3a207816cc569f78863d86c01f2a3d265350e39f 2025-09-07T07:36:04.3691531Z * [new tag] trunk/3a20a20e7065ec927fdd216d4da3b04f879b3c67 -> trunk/3a20a20e7065ec927fdd216d4da3b04f879b3c67 2025-09-07T07:36:04.3692546Z * [new tag] trunk/3bbc2e3e4f025523eaa5dbff220b3e96bca608d0 -> trunk/3bbc2e3e4f025523eaa5dbff220b3e96bca608d0 2025-09-07T07:36:04.3693503Z * [new tag] trunk/3c0ff1b569c45cfa6935ad8031a9d4cf1551aa3f -> trunk/3c0ff1b569c45cfa6935ad8031a9d4cf1551aa3f 2025-09-07T07:36:04.3694466Z * [new tag] trunk/3c45af079afc92a03b03ddf4f9198902ffcf30cf -> trunk/3c45af079afc92a03b03ddf4f9198902ffcf30cf 2025-09-07T07:36:04.3695461Z * [new tag] trunk/3dde5d7f9bf80dd6623a712bc429e9e4302464b5 -> trunk/3dde5d7f9bf80dd6623a712bc429e9e4302464b5 2025-09-07T07:36:04.3697530Z * [new tag] trunk/403a3a393cda7e60f503f3b04b8805a845dcf45d -> trunk/403a3a393cda7e60f503f3b04b8805a845dcf45d 2025-09-07T07:36:04.3698097Z * [new tag] trunk/420c52ecf36f86d32da0853bfbe074b682b070aa -> trunk/420c52ecf36f86d32da0853bfbe074b682b070aa 2025-09-07T07:36:04.3699945Z * [new tag] trunk/43b7c86a2c0f91320f5c5f4827b111edff06fdb6 -> trunk/43b7c86a2c0f91320f5c5f4827b111edff06fdb6 2025-09-07T07:36:04.3700462Z * [new tag] trunk/451ed931562ec8b46d1f7e6c266a68132a119336 -> trunk/451ed931562ec8b46d1f7e6c266a68132a119336 2025-09-07T07:36:04.3700959Z * [new tag] trunk/480c7391126656154318fabf1d57ebc01e196e63 -> trunk/480c7391126656154318fabf1d57ebc01e196e63 2025-09-07T07:36:04.3701469Z * [new tag] trunk/48bedd753da22634aa94fbafeb731e82025404f3 -> trunk/48bedd753da22634aa94fbafeb731e82025404f3 2025-09-07T07:36:04.3702094Z * [new tag] trunk/494878a11b79071ada0b98f34042d47155be6d1c -> trunk/494878a11b79071ada0b98f34042d47155be6d1c 2025-09-07T07:36:04.3703170Z * [new tag] trunk/4ae57d448c0a7d37e4cfd5c27d977fad2cef4051 -> trunk/4ae57d448c0a7d37e4cfd5c27d977fad2cef4051 2025-09-07T07:36:04.3704225Z * [new tag] trunk/4cdaf8265d86f984254b62052da8c26ef61ef1cf -> trunk/4cdaf8265d86f984254b62052da8c26ef61ef1cf 2025-09-07T07:36:04.3705133Z * [new tag] trunk/4d4abec80f03cd8fdefe1d9cb3a60d3690cd777e -> trunk/4d4abec80f03cd8fdefe1d9cb3a60d3690cd777e 2025-09-07T07:36:04.3706154Z * [new tag] trunk/4e42aa8ffc44b8340eb0eeaf80a2cafc4763a186 -> trunk/4e42aa8ffc44b8340eb0eeaf80a2cafc4763a186 2025-09-07T07:36:04.3707087Z * [new tag] trunk/4f72d932feee0749397fec876dcd43994f50b215 -> trunk/4f72d932feee0749397fec876dcd43994f50b215 2025-09-07T07:36:04.3708074Z * [new tag] trunk/50fc22dedf3c4a27be61fa05551c4f320281b42d -> trunk/50fc22dedf3c4a27be61fa05551c4f320281b42d 2025-09-07T07:36:04.3709040Z * [new tag] trunk/5211f1f908907ffc064b56e43cf8659f7fc22aa9 -> trunk/5211f1f908907ffc064b56e43cf8659f7fc22aa9 2025-09-07T07:36:04.3710034Z * [new tag] trunk/524b78d4f67045b83bb69edc56ab16efe282971c -> trunk/524b78d4f67045b83bb69edc56ab16efe282971c 2025-09-07T07:36:04.3711046Z * [new tag] trunk/54e275e0d81fe1e1ccfa4fb5f2a5a9aaca00ca15 -> trunk/54e275e0d81fe1e1ccfa4fb5f2a5a9aaca00ca15 2025-09-07T07:36:04.3711910Z * [new tag] trunk/5561e45758d59c94605873d5db48ed459c004c3b -> trunk/5561e45758d59c94605873d5db48ed459c004c3b 2025-09-07T07:36:04.3712951Z * [new tag] trunk/57278d45f046d4f89f45d373b1af4dd56934ff24 -> trunk/57278d45f046d4f89f45d373b1af4dd56934ff24 2025-09-07T07:36:04.3714018Z * [new tag] trunk/5927a70934ccf7b70182d364c23245a7dd685503 -> trunk/5927a70934ccf7b70182d364c23245a7dd685503 2025-09-07T07:36:04.3714971Z * [new tag] trunk/5985e28912aeb40b103ebfcf2fd0665eb4a50599 -> trunk/5985e28912aeb40b103ebfcf2fd0665eb4a50599 2025-09-07T07:36:04.3715957Z * [new tag] trunk/5a2da090ed6db88bb657c4e51ec0b310cd08bff6 -> trunk/5a2da090ed6db88bb657c4e51ec0b310cd08bff6 2025-09-07T07:36:04.3716905Z * [new tag] trunk/5c473e9f5ee0ef0fc38e6cf34a95b547f8cdc8d5 -> trunk/5c473e9f5ee0ef0fc38e6cf34a95b547f8cdc8d5 2025-09-07T07:36:04.3718153Z * [new tag] trunk/5c67426d6847667a7c55a2dd01f470fa37238c18 -> trunk/5c67426d6847667a7c55a2dd01f470fa37238c18 2025-09-07T07:36:04.3719053Z * [new tag] trunk/5da573c42c332bc68d4b7946c69f690a876d951a -> trunk/5da573c42c332bc68d4b7946c69f690a876d951a 2025-09-07T07:36:04.3720024Z * [new tag] trunk/5e5870e858f60ff4bf87d03f3592097e934a9580 -> trunk/5e5870e858f60ff4bf87d03f3592097e934a9580 2025-09-07T07:36:04.3720992Z * [new tag] trunk/5f3cbc9442aa55b5afb29f4ac8ca9be569003e84 -> trunk/5f3cbc9442aa55b5afb29f4ac8ca9be569003e84 2025-09-07T07:36:04.3722054Z * [new tag] trunk/600c25e9a17fe56e3dee872be8854db08916ba0c -> trunk/600c25e9a17fe56e3dee872be8854db08916ba0c 2025-09-07T07:36:04.3722899Z * [new tag] trunk/601ae8e4831fc8123fffcfb8fd2e6b6381b42e14 -> trunk/601ae8e4831fc8123fffcfb8fd2e6b6381b42e14 2025-09-07T07:36:04.3723922Z * [new tag] trunk/6087ef41e54c2494b117ffd923faf20f515a6806 -> trunk/6087ef41e54c2494b117ffd923faf20f515a6806 2025-09-07T07:36:04.3724878Z * [new tag] trunk/626cb7df8161dd4ecb4fe43b60f37ce9076f56b1 -> trunk/626cb7df8161dd4ecb4fe43b60f37ce9076f56b1 2025-09-07T07:36:04.3725814Z * [new tag] trunk/62c3f9a97fd3dea7132a93066d32d893ffe101e6 -> trunk/62c3f9a97fd3dea7132a93066d32d893ffe101e6 2025-09-07T07:36:04.3726761Z * [new tag] trunk/63a9c23fe99eacfd09610c36dfe8f01b053c1a35 -> trunk/63a9c23fe99eacfd09610c36dfe8f01b053c1a35 2025-09-07T07:36:04.3727724Z * [new tag] trunk/65985937d97505f648b6ed852c3129f2dd08b251 -> trunk/65985937d97505f648b6ed852c3129f2dd08b251 2025-09-07T07:36:04.3729179Z * [new tag] trunk/66f3b4a682a6153517dd23369fdc3289b6494b07 -> trunk/66f3b4a682a6153517dd23369fdc3289b6494b07 2025-09-07T07:36:04.3729941Z * [new tag] trunk/6737e2c996990024187ba620d2764f3b6f6add2c -> trunk/6737e2c996990024187ba620d2764f3b6f6add2c 2025-09-07T07:36:04.3730933Z * [new tag] trunk/67c31dcd364f10072a55f4a30ffd1151c686283a -> trunk/67c31dcd364f10072a55f4a30ffd1151c686283a 2025-09-07T07:36:04.3731952Z * [new tag] trunk/68738beff73e9c3512e18b4edea811a897ce42db -> trunk/68738beff73e9c3512e18b4edea811a897ce42db 2025-09-07T07:36:04.3732973Z * [new tag] trunk/69a25f68884a168550695fdb1a7c310c54d29536 -> trunk/69a25f68884a168550695fdb1a7c310c54d29536 2025-09-07T07:36:04.3733899Z * [new tag] trunk/6b1900c22f1a07b9519346898d4c71d8a2b0f12f -> trunk/6b1900c22f1a07b9519346898d4c71d8a2b0f12f 2025-09-07T07:36:04.3734891Z * [new tag] trunk/6b8b3ac4403f771bd4a8f9a45d93347304148774 -> trunk/6b8b3ac4403f771bd4a8f9a45d93347304148774 2025-09-07T07:36:04.3735786Z * [new tag] trunk/6f7608d603834d6068b2e7a5d59bec3973b6bb1b -> trunk/6f7608d603834d6068b2e7a5d59bec3973b6bb1b 2025-09-07T07:36:04.3736829Z * [new tag] trunk/70d36e047dfb3488fd6335016711a784d810ebda -> trunk/70d36e047dfb3488fd6335016711a784d810ebda 2025-09-07T07:36:04.3737844Z * [new tag] trunk/71992dd805ff9d6763f77214dfe8b0465e88c87b -> trunk/71992dd805ff9d6763f77214dfe8b0465e88c87b 2025-09-07T07:36:04.3738816Z * [new tag] trunk/734ce8eba9c69381f187359bf0fef1d71d84cd20 -> trunk/734ce8eba9c69381f187359bf0fef1d71d84cd20 2025-09-07T07:36:04.3739809Z * [new tag] trunk/73eb4511fb863a37944342b7e92aae706de603c8 -> trunk/73eb4511fb863a37944342b7e92aae706de603c8 2025-09-07T07:36:04.3740809Z * [new tag] trunk/75bc23cfc345bd4c05e7f97c416c4b3d2d1fa64b -> trunk/75bc23cfc345bd4c05e7f97c416c4b3d2d1fa64b 2025-09-07T07:36:04.3741791Z * [new tag] trunk/771f369448321a387f2018535bc8b8b6e5f12fab -> trunk/771f369448321a387f2018535bc8b8b6e5f12fab 2025-09-07T07:36:04.3742842Z * [new tag] trunk/789d4942127143f2adcb53612c058ce4c9a2cf20 -> trunk/789d4942127143f2adcb53612c058ce4c9a2cf20 2025-09-07T07:36:04.3743672Z * [new tag] trunk/791eff96c85678c950888f9da24650083ee673fe -> trunk/791eff96c85678c950888f9da24650083ee673fe 2025-09-07T07:36:04.3744518Z * [new tag] trunk/793fc12aff1f69fbbf9f4278182fb52bbe350fc9 -> trunk/793fc12aff1f69fbbf9f4278182fb52bbe350fc9 2025-09-07T07:36:04.3745440Z * [new tag] trunk/79fcd5247a9a129eee526a14df30bfc6a22b3f01 -> trunk/79fcd5247a9a129eee526a14df30bfc6a22b3f01 2025-09-07T07:36:04.3746385Z * [new tag] trunk/7f4ff79210eb06924f223ae3a1941ee0e2635348 -> trunk/7f4ff79210eb06924f223ae3a1941ee0e2635348 2025-09-07T07:36:04.3747402Z * [new tag] trunk/8076a185c85112be62be292eb47409c88a585b1c -> trunk/8076a185c85112be62be292eb47409c88a585b1c 2025-09-07T07:36:04.3748300Z * [new tag] trunk/80dd397f1979371a5583fa3d5c7352029522a78d -> trunk/80dd397f1979371a5583fa3d5c7352029522a78d 2025-09-07T07:36:04.3749152Z * [new tag] trunk/8171d6052ec12628eb67e0040839314056014429 -> trunk/8171d6052ec12628eb67e0040839314056014429 2025-09-07T07:36:04.3750595Z * [new tag] trunk/81aeefa657b7ccc26b275c50a9f33b2f056e8071 -> trunk/81aeefa657b7ccc26b275c50a9f33b2f056e8071 2025-09-07T07:36:04.3751518Z * [new tag] trunk/81b7b16618bda250ce55982894a83dc0805eb64c -> trunk/81b7b16618bda250ce55982894a83dc0805eb64c 2025-09-07T07:36:04.3752511Z * [new tag] trunk/827f0d405448de31f79d1089f7d7fceab2f87895 -> trunk/827f0d405448de31f79d1089f7d7fceab2f87895 2025-09-07T07:36:04.3753505Z * [new tag] trunk/82f63c8f6de63c30132a8ac299b6e8c2fd0d3fe8 -> trunk/82f63c8f6de63c30132a8ac299b6e8c2fd0d3fe8 2025-09-07T07:36:04.3754473Z * [new tag] trunk/850e1382a9c56bfde18af09d3e72352d775e9435 -> trunk/850e1382a9c56bfde18af09d3e72352d775e9435 2025-09-07T07:36:04.3755486Z * [new tag] trunk/8678d831c48e616b717bff50f2d03141d2e9f965 -> trunk/8678d831c48e616b717bff50f2d03141d2e9f965 2025-09-07T07:36:04.3756518Z * [new tag] trunk/869cbcc16e489a4f5a14a93d5779b0ea86061c60 -> trunk/869cbcc16e489a4f5a14a93d5779b0ea86061c60 2025-09-07T07:36:04.3757527Z * [new tag] trunk/8703debf669bc2238211bfd039f4ecdd8228b7f7 -> trunk/8703debf669bc2238211bfd039f4ecdd8228b7f7 2025-09-07T07:36:04.3758495Z * [new tag] trunk/874069fbe46e82da5cfa405e6c0deb12e89ff608 -> trunk/874069fbe46e82da5cfa405e6c0deb12e89ff608 2025-09-07T07:36:04.3759544Z * [new tag] trunk/8875d6e394da2fffd04f31b28bf258c94d4776a3 -> trunk/8875d6e394da2fffd04f31b28bf258c94d4776a3 2025-09-07T07:36:04.3760542Z * [new tag] trunk/88d94d17e8c5155451393afa6eb3bab48ab61c16 -> trunk/88d94d17e8c5155451393afa6eb3bab48ab61c16 2025-09-07T07:36:04.3761549Z * [new tag] trunk/890626632def7e0ef95a2d01e87a0e4627824a9f -> trunk/890626632def7e0ef95a2d01e87a0e4627824a9f 2025-09-07T07:36:04.3762608Z * [new tag] trunk/8975cda2520b7b1b5bc3b4d8213edf261fa82570 -> trunk/8975cda2520b7b1b5bc3b4d8213edf261fa82570 2025-09-07T07:36:04.3763626Z * [new tag] trunk/89d41d3f61d04f14730ec26f008a59bef6624610 -> trunk/89d41d3f61d04f14730ec26f008a59bef6624610 2025-09-07T07:36:04.3764626Z * [new tag] trunk/8bb213b6d599ef1273fe52f9b1f6d476056c3a41 -> trunk/8bb213b6d599ef1273fe52f9b1f6d476056c3a41 2025-09-07T07:36:04.3765571Z * [new tag] trunk/8e23a1227b5fb2e39afaa7d57c075a75b640a5af -> trunk/8e23a1227b5fb2e39afaa7d57c075a75b640a5af 2025-09-07T07:36:04.3767062Z * [new tag] trunk/8ec551bb354ab2b85fbbba9d461740a20366d248 -> trunk/8ec551bb354ab2b85fbbba9d461740a20366d248 2025-09-07T07:36:04.3768048Z * [new tag] trunk/8fd3c9ce919c8d5c645fd348bba517e948cbc29d -> trunk/8fd3c9ce919c8d5c645fd348bba517e948cbc29d 2025-09-07T07:36:04.3769360Z * [new tag] trunk/90f50f7e68e120d9574e6e3189e37b4280010ad9 -> trunk/90f50f7e68e120d9574e6e3189e37b4280010ad9 2025-09-07T07:36:04.3770356Z * [new tag] trunk/91f0bcf43fc0bc743350d491ac63b77e92054ac9 -> trunk/91f0bcf43fc0bc743350d491ac63b77e92054ac9 2025-09-07T07:36:04.3771447Z * [new tag] trunk/92576a594b8121f6b0b1b5a3ea16d08792fc68ab -> trunk/92576a594b8121f6b0b1b5a3ea16d08792fc68ab 2025-09-07T07:36:04.3772497Z * [new tag] trunk/92a43025e0baa1f2ce345f28d22913b518a1ab9d -> trunk/92a43025e0baa1f2ce345f28d22913b518a1ab9d 2025-09-07T07:36:04.3773359Z * [new tag] trunk/93fb23d6fae7c4e82c4239a1033e522088742634 -> trunk/93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:36:04.3774364Z * [new tag] trunk/9458d1ac3bd70c2af316a8ba95d2c6c9c1199c9c -> trunk/9458d1ac3bd70c2af316a8ba95d2c6c9c1199c9c 2025-09-07T07:36:04.3775465Z * [new tag] trunk/9480cdc0b61488c89a23c2f64f43b2dcedc8728e -> trunk/9480cdc0b61488c89a23c2f64f43b2dcedc8728e 2025-09-07T07:36:04.3776721Z * [new tag] trunk/9491d289b329e4ba4a9f5f5b1be7960671bb7840 -> trunk/9491d289b329e4ba4a9f5f5b1be7960671bb7840 2025-09-07T07:36:04.3777635Z * [new tag] trunk/9499c8761cd2067feb9877414e818f6fd00290f1 -> trunk/9499c8761cd2067feb9877414e818f6fd00290f1 2025-09-07T07:36:04.3778934Z * [new tag] trunk/95ee0bfea99d3d346d6502b91b497d2b35795504 -> trunk/95ee0bfea99d3d346d6502b91b497d2b35795504 2025-09-07T07:36:04.3779898Z * [new tag] trunk/98374612fc2febd686be20761e56bdc2424bc36a -> trunk/98374612fc2febd686be20761e56bdc2424bc36a 2025-09-07T07:36:04.3780965Z * [new tag] trunk/98efc9e93d8fc61eb53cb91378443617cb550500 -> trunk/98efc9e93d8fc61eb53cb91378443617cb550500 2025-09-07T07:36:04.3781994Z * [new tag] trunk/994f2a5dbcbdc915da39bf6f6ce4d1f5e74835c9 -> trunk/994f2a5dbcbdc915da39bf6f6ce4d1f5e74835c9 2025-09-07T07:36:04.3782998Z * [new tag] trunk/99f356fa58c8d726cef022d8710f5491291158f6 -> trunk/99f356fa58c8d726cef022d8710f5491291158f6 2025-09-07T07:36:04.3783987Z * [new tag] trunk/9a1c5c0a078b94d13ac5c1ae0d754d19fb73bf99 -> trunk/9a1c5c0a078b94d13ac5c1ae0d754d19fb73bf99 2025-09-07T07:36:04.3784972Z * [new tag] trunk/9a665ca3c472384e9d722bddba79e5a7680f1abd -> trunk/9a665ca3c472384e9d722bddba79e5a7680f1abd 2025-09-07T07:36:04.3785984Z * [new tag] trunk/9aedb3cd87b52160872173c177f61053d97bed57 -> trunk/9aedb3cd87b52160872173c177f61053d97bed57 2025-09-07T07:36:04.3786948Z * [new tag] trunk/9b81fe281da41f2421506339d26b027a468902f4 -> trunk/9b81fe281da41f2421506339d26b027a468902f4 2025-09-07T07:36:04.3787966Z * [new tag] trunk/9bdcee01f86e2969cff1140cdecfca13cb51816e -> trunk/9bdcee01f86e2969cff1140cdecfca13cb51816e 2025-09-07T07:36:04.3788994Z * [new tag] trunk/9c03d6be87eedc06e524e202e07a7e776551a839 -> trunk/9c03d6be87eedc06e524e202e07a7e776551a839 2025-09-07T07:36:04.3790035Z * [new tag] trunk/9c957723a0fedd9c637e63e023a613019e2cab60 -> trunk/9c957723a0fedd9c637e63e023a613019e2cab60 2025-09-07T07:36:04.3791009Z * [new tag] trunk/9e5247f51d81735e5f1e65e80588985fa93bccc5 -> trunk/9e5247f51d81735e5f1e65e80588985fa93bccc5 2025-09-07T07:36:04.3792043Z * [new tag] trunk/9eadb37cdd699f7e8e8177a5227bfeb16184ef26 -> trunk/9eadb37cdd699f7e8e8177a5227bfeb16184ef26 2025-09-07T07:36:04.3793037Z * [new tag] trunk/a00cdc1e4159db73c9ffb3f25e93e55877709a29 -> trunk/a00cdc1e4159db73c9ffb3f25e93e55877709a29 2025-09-07T07:36:04.3794027Z * [new tag] trunk/a02ee4a816d11380c6f564c1aba64d56af5ba705 -> trunk/a02ee4a816d11380c6f564c1aba64d56af5ba705 2025-09-07T07:36:04.3795008Z * [new tag] trunk/a3c7f77e50f900721817934120d60c2361b3c40d -> trunk/a3c7f77e50f900721817934120d60c2361b3c40d 2025-09-07T07:36:04.3796051Z * [new tag] trunk/a3d72b09ae12126a2b7d4a63a45ac100a882a802 -> trunk/a3d72b09ae12126a2b7d4a63a45ac100a882a802 2025-09-07T07:36:04.3797034Z * [new tag] trunk/a3e5466002791da609fcb069155d8ee347baee92 -> trunk/a3e5466002791da609fcb069155d8ee347baee92 2025-09-07T07:36:04.3798025Z * [new tag] trunk/a714437093ed196eee28f7de454cf4c41badc098 -> trunk/a714437093ed196eee28f7de454cf4c41badc098 2025-09-07T07:36:04.3798983Z * [new tag] trunk/a75e8cd27098f290de0b7439685d05ce02e91356 -> trunk/a75e8cd27098f290de0b7439685d05ce02e91356 2025-09-07T07:36:04.3800007Z * [new tag] trunk/a8d6943d36c1c2a5f90d3573460695bad4b623ae -> trunk/a8d6943d36c1c2a5f90d3573460695bad4b623ae 2025-09-07T07:36:04.3800818Z * [new tag] trunk/a918bbad6ab20649ff82eefb48417ecbe96bcb34 -> trunk/a918bbad6ab20649ff82eefb48417ecbe96bcb34 2025-09-07T07:36:04.3802015Z * [new tag] trunk/a99d8d39bc842d6ebc3e368b178e4884d24b056e -> trunk/a99d8d39bc842d6ebc3e368b178e4884d24b056e 2025-09-07T07:36:04.3802934Z * [new tag] trunk/aac1a50a191b4102d566c9c1ea22f06d6c2e3f02 -> trunk/aac1a50a191b4102d566c9c1ea22f06d6c2e3f02 2025-09-07T07:36:04.3804160Z * [new tag] trunk/aad96a202244c7d0d120c04ba8db593edd8c0f92 -> trunk/aad96a202244c7d0d120c04ba8db593edd8c0f92 2025-09-07T07:36:04.3805146Z * [new tag] trunk/ab643e4dbbaf7b663d4237514cbf01af9b11565c -> trunk/ab643e4dbbaf7b663d4237514cbf01af9b11565c 2025-09-07T07:36:04.3806173Z * [new tag] trunk/abc447174cd2cf8591edbc70a9f836f9a5779f47 -> trunk/abc447174cd2cf8591edbc70a9f836f9a5779f47 2025-09-07T07:36:04.3807126Z * [new tag] trunk/acece97c3a9dceb63194e314da93fdf37cf15a0d -> trunk/acece97c3a9dceb63194e314da93fdf37cf15a0d 2025-09-07T07:36:04.3808230Z * [new tag] trunk/adae7f66aacf3f248c3101b858cf98d5809119fa -> trunk/adae7f66aacf3f248c3101b858cf98d5809119fa 2025-09-07T07:36:04.3809218Z * [new tag] trunk/ae0edc133e61e3b16caf0b2ee0ff3f33ab72af4c -> trunk/ae0edc133e61e3b16caf0b2ee0ff3f33ab72af4c 2025-09-07T07:36:04.3810171Z * [new tag] trunk/aed33a8fcbd60b052d4559d261390c5797129c6d -> trunk/aed33a8fcbd60b052d4559d261390c5797129c6d 2025-09-07T07:36:04.3811179Z * [new tag] trunk/b04e922712080a3652e438d05e8bb74e0cd2d238 -> trunk/b04e922712080a3652e438d05e8bb74e0cd2d238 2025-09-07T07:36:04.3812222Z * [new tag] trunk/b0a3e58dd71c1a039ac0ef51e5bd8f704f632f6f -> trunk/b0a3e58dd71c1a039ac0ef51e5bd8f704f632f6f 2025-09-07T07:36:04.3813219Z * [new tag] trunk/b16d3f4c8c01d461c2f01064e9ca5fa2b33f5cf1 -> trunk/b16d3f4c8c01d461c2f01064e9ca5fa2b33f5cf1 2025-09-07T07:36:04.3814178Z * [new tag] trunk/b18bb6796f210a183e687d9d64984a5a9d13cf09 -> trunk/b18bb6796f210a183e687d9d64984a5a9d13cf09 2025-09-07T07:36:04.3815124Z * [new tag] trunk/b1bb98ddebdd3e41bf7987372409bdce96ae55de -> trunk/b1bb98ddebdd3e41bf7987372409bdce96ae55de 2025-09-07T07:36:04.3816679Z * [new tag] trunk/b2b4add0e754411372060e1d7b4057a66439172b -> trunk/b2b4add0e754411372060e1d7b4057a66439172b 2025-09-07T07:36:04.3817794Z * [new tag] trunk/b2c7b9ad2dc5a7c0b61febd307761bd5bc2f0f05 -> trunk/b2c7b9ad2dc5a7c0b61febd307761bd5bc2f0f05 2025-09-07T07:36:04.3819000Z * [new tag] trunk/b40d9432be44a6b5974ee62e7d19c3c61c5ece37 -> trunk/b40d9432be44a6b5974ee62e7d19c3c61c5ece37 2025-09-07T07:36:04.3820018Z * [new tag] trunk/b4ad38279b178b7bd14355123c1101e2e853e77b -> trunk/b4ad38279b178b7bd14355123c1101e2e853e77b 2025-09-07T07:36:04.3821048Z * [new tag] trunk/b67c41039835bd9b20b83cd6233e86baaa5f5dde -> trunk/b67c41039835bd9b20b83cd6233e86baaa5f5dde 2025-09-07T07:36:04.3822149Z * [new tag] trunk/b6d0a9ea9056ede4f7024dbf3bd6c43be3aff49c -> trunk/b6d0a9ea9056ede4f7024dbf3bd6c43be3aff49c 2025-09-07T07:36:04.3823147Z * [new tag] trunk/b7dad7dd49448c88d0751fa2e29c70afe985f734 -> trunk/b7dad7dd49448c88d0751fa2e29c70afe985f734 2025-09-07T07:36:04.3824139Z * [new tag] trunk/b7e207ca9f046ddd716076965a0cce403ba99052 -> trunk/b7e207ca9f046ddd716076965a0cce403ba99052 2025-09-07T07:36:04.3825122Z * [new tag] trunk/b919560c4a7010e2d89facee25586269a994746e -> trunk/b919560c4a7010e2d89facee25586269a994746e 2025-09-07T07:36:04.3826151Z * [new tag] trunk/b9ba612f7a968f7b27e121ca8f4d0a4d954f5354 -> trunk/b9ba612f7a968f7b27e121ca8f4d0a4d954f5354 2025-09-07T07:36:04.3827213Z * [new tag] trunk/ba7f546ccccb5e0b36d9070dc25f26a9647f89f8 -> trunk/ba7f546ccccb5e0b36d9070dc25f26a9647f89f8 2025-09-07T07:36:04.3828211Z * [new tag] trunk/bb950284c7e72905994bc25dd436c10e48088d85 -> trunk/bb950284c7e72905994bc25dd436c10e48088d85 2025-09-07T07:36:04.3829200Z * [new tag] trunk/bbedc71fd3267c639c38b4ec25eaa22f973d9c4d -> trunk/bbedc71fd3267c639c38b4ec25eaa22f973d9c4d 2025-09-07T07:36:04.3830111Z * [new tag] trunk/bc4db2c27fce6ff1648bdc5af31ec225d2a31f37 -> trunk/bc4db2c27fce6ff1648bdc5af31ec225d2a31f37 2025-09-07T07:36:04.3831043Z * [new tag] trunk/bc505977fb66677a09c31155c987330fbb18a865 -> trunk/bc505977fb66677a09c31155c987330fbb18a865 2025-09-07T07:36:04.3832027Z * [new tag] trunk/bd39e47feea7326afb5bbb67fcb1e69279239527 -> trunk/bd39e47feea7326afb5bbb67fcb1e69279239527 2025-09-07T07:36:04.3833125Z * [new tag] trunk/be5b03dde96638f25ffd732a4fed7e41b4cf40e1 -> trunk/be5b03dde96638f25ffd732a4fed7e41b4cf40e1 2025-09-07T07:36:04.3834117Z * [new tag] trunk/bffc7dd1f374d8408911cd22c6b3d6df39ded9b3 -> trunk/bffc7dd1f374d8408911cd22c6b3d6df39ded9b3 2025-09-07T07:36:04.3835208Z * [new tag] trunk/c024b1f5a18d5c5aee5cc2acdd4c52b24b93ffcf -> trunk/c024b1f5a18d5c5aee5cc2acdd4c52b24b93ffcf 2025-09-07T07:36:04.3836190Z * [new tag] trunk/c0983e6cc0acf71689e1851d12609e00b3f59371 -> trunk/c0983e6cc0acf71689e1851d12609e00b3f59371 2025-09-07T07:36:04.3837169Z * [new tag] trunk/c10195e723eeeedd099ed8b73eda7184ca618fad -> trunk/c10195e723eeeedd099ed8b73eda7184ca618fad 2025-09-07T07:36:04.3838186Z * [new tag] trunk/c157cf6488ade6a7ee2ce2d25b059e1335630a99 -> trunk/c157cf6488ade6a7ee2ce2d25b059e1335630a99 2025-09-07T07:36:04.3839190Z * [new tag] trunk/c2a30246172fd71d56529907ffd3c27b76b1f3a7 -> trunk/c2a30246172fd71d56529907ffd3c27b76b1f3a7 2025-09-07T07:36:04.3840181Z * [new tag] trunk/c32111149921b48bfef909293f1049e21619ed76 -> trunk/c32111149921b48bfef909293f1049e21619ed76 2025-09-07T07:36:04.3840997Z * [new tag] trunk/c37103234afc832dcad307e9016230810957c9d5 -> trunk/c37103234afc832dcad307e9016230810957c9d5 2025-09-07T07:36:04.3842035Z * [new tag] trunk/c3ceca2995cd35e1376c4b0704669bff1a81e836 -> trunk/c3ceca2995cd35e1376c4b0704669bff1a81e836 2025-09-07T07:36:04.3843041Z * [new tag] trunk/c3d54dea9febb1236d48d19e5d4876a63f2e20fd -> trunk/c3d54dea9febb1236d48d19e5d4876a63f2e20fd 2025-09-07T07:36:04.3844132Z * [new tag] trunk/c465b3d52c5687fe910d35a5c75341b77f821741 -> trunk/c465b3d52c5687fe910d35a5c75341b77f821741 2025-09-07T07:36:04.3845150Z * [new tag] trunk/c5b8a10be5e89396da916d1069ffcb7135f0372b -> trunk/c5b8a10be5e89396da916d1069ffcb7135f0372b 2025-09-07T07:36:04.3846105Z * [new tag] trunk/c7e41071a08f4045bc11ab60ec366d7357d56e30 -> trunk/c7e41071a08f4045bc11ab60ec366d7357d56e30 2025-09-07T07:36:04.3847051Z * [new tag] trunk/c98ddaca6d2e19ca37aff00c4ff0cda1e9a6ff65 -> trunk/c98ddaca6d2e19ca37aff00c4ff0cda1e9a6ff65 2025-09-07T07:36:04.3848010Z * [new tag] trunk/cb1e31362c7b53acf4ac95b9f8878064c184f03b -> trunk/cb1e31362c7b53acf4ac95b9f8878064c184f03b 2025-09-07T07:36:04.3848993Z * [new tag] trunk/cbfb005f7cce79974795b148e265f594f59477c8 -> trunk/cbfb005f7cce79974795b148e265f594f59477c8 2025-09-07T07:36:04.3850057Z * [new tag] trunk/cc5bdd12401bda835291d2f3cb297132ebdbf358 -> trunk/cc5bdd12401bda835291d2f3cb297132ebdbf358 2025-09-07T07:36:04.3851163Z * [new tag] trunk/cd529b686d54bbaa443f5b310140de48422d96c7 -> trunk/cd529b686d54bbaa443f5b310140de48422d96c7 2025-09-07T07:36:04.3852119Z * [new tag] trunk/cec0ff122815582af5302360aff03676558c5c87 -> trunk/cec0ff122815582af5302360aff03676558c5c87 2025-09-07T07:36:04.3853126Z * [new tag] trunk/d11720efdb563d02cf4f7d324311fb15a755268e -> trunk/d11720efdb563d02cf4f7d324311fb15a755268e 2025-09-07T07:36:04.3854107Z * [new tag] trunk/d1706d9128ae24d9048167e80d3fe5196d19035e -> trunk/d1706d9128ae24d9048167e80d3fe5196d19035e 2025-09-07T07:36:04.3855113Z * [new tag] trunk/d1a15abfdcaef138f2d9e93a9f46be44f30b766d -> trunk/d1a15abfdcaef138f2d9e93a9f46be44f30b766d 2025-09-07T07:36:04.3856316Z * [new tag] trunk/d232a95d4a79404ca05c1f52d37fde7339dcdf49 -> trunk/d232a95d4a79404ca05c1f52d37fde7339dcdf49 2025-09-07T07:36:04.3857289Z * [new tag] trunk/d2d4c8e9b2371c9aacfb771d9402ac7427b9778e -> trunk/d2d4c8e9b2371c9aacfb771d9402ac7427b9778e 2025-09-07T07:36:04.3858353Z * [new tag] trunk/d33840c542b387ab08ba49aa6c45aa9567fd9be7 -> trunk/d33840c542b387ab08ba49aa6c45aa9567fd9be7 2025-09-07T07:36:04.3859330Z * [new tag] trunk/d5643e8f3a648a99636bfa1f2a41d54bd3c0d0f1 -> trunk/d5643e8f3a648a99636bfa1f2a41d54bd3c0d0f1 2025-09-07T07:36:04.3860295Z * [new tag] trunk/d5b38410b5b6cf75c7a7389972777a6497926ee7 -> trunk/d5b38410b5b6cf75c7a7389972777a6497926ee7 2025-09-07T07:36:04.3861231Z * [new tag] trunk/d5e0f4202ba14632e4d14862ace096609e763462 -> trunk/d5e0f4202ba14632e4d14862ace096609e763462 2025-09-07T07:36:04.3862220Z * [new tag] trunk/d636c181f9140a7b59be10b36eae23039fc2bb72 -> trunk/d636c181f9140a7b59be10b36eae23039fc2bb72 2025-09-07T07:36:04.3863607Z * [new tag] trunk/d64718503728001a1e78168fd7f2d4ff23e57285 -> trunk/d64718503728001a1e78168fd7f2d4ff23e57285 2025-09-07T07:36:04.3864660Z * [new tag] trunk/d67c29ad22670320d676b02e394274af34e8e643 -> trunk/d67c29ad22670320d676b02e394274af34e8e643 2025-09-07T07:36:04.3865653Z * [new tag] trunk/d6b74568e2c98ce58ecc145b72ac66d4caf7ce95 -> trunk/d6b74568e2c98ce58ecc145b72ac66d4caf7ce95 2025-09-07T07:36:04.3866722Z * [new tag] trunk/d711f27845abd45007ccab6076649ebd896c2661 -> trunk/d711f27845abd45007ccab6076649ebd896c2661 2025-09-07T07:36:04.3867797Z * [new tag] trunk/d9d6dde0f42d4bcc8c97671ac50d5096c7e500ab -> trunk/d9d6dde0f42d4bcc8c97671ac50d5096c7e500ab 2025-09-07T07:36:04.3869001Z * [new tag] trunk/da4db4b33d1fdd046650cf19fdbac581a19bf2f9 -> trunk/da4db4b33d1fdd046650cf19fdbac581a19bf2f9 2025-09-07T07:36:04.3869846Z * [new tag] trunk/dac8a4b91c01c3bbc96f54e621b1ea4ffdbd29d1 -> trunk/dac8a4b91c01c3bbc96f54e621b1ea4ffdbd29d1 2025-09-07T07:36:04.3870974Z * [new tag] trunk/dbec08729fb9848bebed6048c63831b87170d061 -> trunk/dbec08729fb9848bebed6048c63831b87170d061 2025-09-07T07:36:04.3871806Z * [new tag] trunk/dcf385395d838f38c8dca25913578230dd43099a -> trunk/dcf385395d838f38c8dca25913578230dd43099a 2025-09-07T07:36:04.3872835Z * [new tag] trunk/dd2519abe83ec3c40d4797492434e41fe3b47e17 -> trunk/dd2519abe83ec3c40d4797492434e41fe3b47e17 2025-09-07T07:36:04.3873829Z * [new tag] trunk/dec72ea4b006dd0fbcaaaa106ad273d73807ab9d -> trunk/dec72ea4b006dd0fbcaaaa106ad273d73807ab9d 2025-09-07T07:36:04.3874800Z * [new tag] trunk/e0a62b266c021b910ce6dc02a6c9429210487717 -> trunk/e0a62b266c021b910ce6dc02a6c9429210487717 2025-09-07T07:36:04.3875835Z * [new tag] trunk/e19e02c84c9dcc408375e5cae3b0709c18b99228 -> trunk/e19e02c84c9dcc408375e5cae3b0709c18b99228 2025-09-07T07:36:04.3876923Z * [new tag] trunk/e304ea4e69d3a7deeb7e48c7450c214a4c953937 -> trunk/e304ea4e69d3a7deeb7e48c7450c214a4c953937 2025-09-07T07:36:04.3877943Z * [new tag] trunk/e3068cdb446adefb5a875616ba37a60235391439 -> trunk/e3068cdb446adefb5a875616ba37a60235391439 2025-09-07T07:36:04.3878933Z * [new tag] trunk/e381d4b0205d5f126c1de534f867ba776f7c3ee6 -> trunk/e381d4b0205d5f126c1de534f867ba776f7c3ee6 2025-09-07T07:36:04.3879988Z * [new tag] trunk/e4bd0ff4f8981b805df32ea5b3550621965ea4f2 -> trunk/e4bd0ff4f8981b805df32ea5b3550621965ea4f2 2025-09-07T07:36:04.3880972Z * [new tag] trunk/e532c9d4f1cdcbc1ea9628f55b9813e77847bdc7 -> trunk/e532c9d4f1cdcbc1ea9628f55b9813e77847bdc7 2025-09-07T07:36:04.3882551Z * [new tag] trunk/e92cd9415377403b6e90585e764639e2e0b5973b -> trunk/e92cd9415377403b6e90585e764639e2e0b5973b 2025-09-07T07:36:04.3883751Z * [new tag] trunk/e9481b6617b5576b099d8ca5798111592e9ad090 -> trunk/e9481b6617b5576b099d8ca5798111592e9ad090 2025-09-07T07:36:04.3884499Z * [new tag] trunk/ea1883dfd3e42defe37b11202b878bb76defa087 -> trunk/ea1883dfd3e42defe37b11202b878bb76defa087 2025-09-07T07:36:04.3885628Z * [new tag] trunk/eac3d6f04cfbbebe3d470dacd216da7d4b1f95a8 -> trunk/eac3d6f04cfbbebe3d470dacd216da7d4b1f95a8 2025-09-07T07:36:04.3886568Z * [new tag] trunk/eb18d32bda75189494d955aa001ade15f10333de -> trunk/eb18d32bda75189494d955aa001ade15f10333de 2025-09-07T07:36:04.3887386Z * [new tag] trunk/ef3be6726f7ff4b77c22db10cec5b686f9107ea9 -> trunk/ef3be6726f7ff4b77c22db10cec5b686f9107ea9 2025-09-07T07:36:04.3888419Z * [new tag] trunk/ef8aabd42422725026cb4dbf48aafa9efa226a04 -> trunk/ef8aabd42422725026cb4dbf48aafa9efa226a04 2025-09-07T07:36:04.3889506Z * [new tag] trunk/f00445b43eee57e20bb9316fa796ca23bf73373b -> trunk/f00445b43eee57e20bb9316fa796ca23bf73373b 2025-09-07T07:36:04.3890534Z * [new tag] trunk/f0c391102b754e3b145e8c59231d2df563487e37 -> trunk/f0c391102b754e3b145e8c59231d2df563487e37 2025-09-07T07:36:04.3891624Z * [new tag] trunk/f27985b7e796fb66a1b476284ba42d8cb360a751 -> trunk/f27985b7e796fb66a1b476284ba42d8cb360a751 2025-09-07T07:36:04.3892692Z * [new tag] trunk/f36f285953700f971552083a5da9d0ceacb63bbd -> trunk/f36f285953700f971552083a5da9d0ceacb63bbd 2025-09-07T07:36:04.3893678Z * [new tag] trunk/f3cebec39ebc110e1c8b06e741896585f7892dbb -> trunk/f3cebec39ebc110e1c8b06e741896585f7892dbb 2025-09-07T07:36:04.3894505Z * [new tag] trunk/f4c33cd44acac92c0b451a04da20ebe9370e5b0c -> trunk/f4c33cd44acac92c0b451a04da20ebe9370e5b0c 2025-09-07T07:36:04.3895612Z * [new tag] trunk/f612045ce105f008b2b675e2fc870163babeb2e8 -> trunk/f612045ce105f008b2b675e2fc870163babeb2e8 2025-09-07T07:36:04.3896619Z * [new tag] trunk/f8746b878dfc1e9639d42cbde832e9b9e792c86c -> trunk/f8746b878dfc1e9639d42cbde832e9b9e792c86c 2025-09-07T07:36:04.3897679Z * [new tag] trunk/f8ffa9194e26523e5f976d4a824d5cc58922727c -> trunk/f8ffa9194e26523e5f976d4a824d5cc58922727c 2025-09-07T07:36:04.3898664Z * [new tag] trunk/f981a7fa5230b98974291fdde32fe8488bc5d469 -> trunk/f981a7fa5230b98974291fdde32fe8488bc5d469 2025-09-07T07:36:04.3899696Z * [new tag] trunk/fbf3d2027daabbcb44d0af274b139be2a248a4f7 -> trunk/fbf3d2027daabbcb44d0af274b139be2a248a4f7 2025-09-07T07:36:04.3900994Z * [new tag] trunk/fca2601c9d628e1bd2d75c7318cd22c4e8c832aa -> trunk/fca2601c9d628e1bd2d75c7318cd22c4e8c832aa 2025-09-07T07:36:04.3902028Z * [new tag] trunk/fea20775ad96bdca972a1811d7d3372f368614ab -> trunk/fea20775ad96bdca972a1811d7d3372f368614ab 2025-09-07T07:36:04.3902861Z * [new tag] trunk/fefee081642f87419a21dc852f7167d4640443cd -> trunk/fefee081642f87419a21dc852f7167d4640443cd 2025-09-07T07:36:04.3904018Z * [new tag] v0.1.1 -> v0.1.1 2025-09-07T07:36:04.3905052Z * [new tag] v0.1.10 -> v0.1.10 2025-09-07T07:36:04.3905909Z * [new tag] v0.1.11 -> v0.1.11 2025-09-07T07:36:04.3906820Z * [new tag] v0.1.12 -> v0.1.12 2025-09-07T07:36:04.3907694Z * [new tag] v0.1.2 -> v0.1.2 2025-09-07T07:36:04.3908567Z * [new tag] v0.1.3 -> v0.1.3 2025-09-07T07:36:04.3909476Z * [new tag] v0.1.4 -> v0.1.4 2025-09-07T07:36:04.3910383Z * [new tag] v0.1.5 -> v0.1.5 2025-09-07T07:36:04.3911283Z * [new tag] v0.1.6 -> v0.1.6 2025-09-07T07:36:04.3912144Z * [new tag] v0.1.7 -> v0.1.7 2025-09-07T07:36:04.3912993Z * [new tag] v0.1.8 -> v0.1.8 2025-09-07T07:36:04.3913968Z * [new tag] v0.1.9 -> v0.1.9 2025-09-07T07:36:04.3914833Z * [new tag] v0.2.0 -> v0.2.0 2025-09-07T07:36:04.3915767Z * [new tag] v0.3.0 -> v0.3.0 2025-09-07T07:36:04.3916781Z * [new tag] v0.3.1 -> v0.3.1 2025-09-07T07:36:04.3917794Z * [new tag] v0.4.0 -> v0.4.0 2025-09-07T07:36:04.3918911Z * [new tag] v0.4.1 -> v0.4.1 2025-09-07T07:36:04.3919858Z * [new tag] v1.0.0 -> v1.0.0 2025-09-07T07:36:04.3920833Z * [new tag] v1.0.0a0 -> v1.0.0a0 2025-09-07T07:36:04.3921746Z * [new tag] v1.0.1 -> v1.0.1 2025-09-07T07:36:04.3922711Z * [new tag] v1.0rc0 -> v1.0rc0 2025-09-07T07:36:04.3923471Z * [new tag] v1.0rc1 -> v1.0rc1 2025-09-07T07:36:04.3924408Z * [new tag] v1.1.0 -> v1.1.0 2025-09-07T07:36:04.3925363Z * [new tag] v1.1.0a0 -> v1.1.0a0 2025-09-07T07:36:04.3926474Z * [new tag] v1.10.0 -> v1.10.0 2025-09-07T07:36:04.3927500Z * [new tag] v1.10.0-rc1 -> v1.10.0-rc1 2025-09-07T07:36:04.3928454Z * [new tag] v1.10.0-rc2 -> v1.10.0-rc2 2025-09-07T07:36:04.3929653Z * [new tag] v1.10.0-rc3 -> v1.10.0-rc3 2025-09-07T07:36:04.3930630Z * [new tag] v1.10.1 -> v1.10.1 2025-09-07T07:36:04.3931448Z * [new tag] v1.10.1-rc1 -> v1.10.1-rc1 2025-09-07T07:36:04.3932218Z * [new tag] v1.10.2 -> v1.10.2 2025-09-07T07:36:04.3932905Z * [new tag] v1.10.2-rc1 -> v1.10.2-rc1 2025-09-07T07:36:04.3933925Z * [new tag] v1.11.0 -> v1.11.0 2025-09-07T07:36:04.3934959Z * [new tag] v1.11.0-rc1 -> v1.11.0-rc1 2025-09-07T07:36:04.3936001Z * [new tag] v1.11.0-rc2 -> v1.11.0-rc2 2025-09-07T07:36:04.3936954Z * [new tag] v1.11.0-rc3 -> v1.11.0-rc3 2025-09-07T07:36:04.3937975Z * [new tag] v1.11.0-rc4 -> v1.11.0-rc4 2025-09-07T07:36:04.3938983Z * [new tag] v1.11.0-rc5 -> v1.11.0-rc5 2025-09-07T07:36:04.3939808Z * [new tag] v1.11.0-rc6 -> v1.11.0-rc6 2025-09-07T07:36:04.3940454Z * [new tag] v1.11.0-rc7 -> v1.11.0-rc7 2025-09-07T07:36:04.3941475Z * [new tag] v1.12.0 -> v1.12.0 2025-09-07T07:36:04.3942458Z * [new tag] v1.12.0-rc1 -> v1.12.0-rc1 2025-09-07T07:36:04.3943404Z * [new tag] v1.12.0-rc2 -> v1.12.0-rc2 2025-09-07T07:36:04.3944338Z * [new tag] v1.12.0-rc3 -> v1.12.0-rc3 2025-09-07T07:36:04.3945303Z * [new tag] v1.12.0-rc4 -> v1.12.0-rc4 2025-09-07T07:36:04.3946392Z * [new tag] v1.12.0-rc5 -> v1.12.0-rc5 2025-09-07T07:36:04.3947365Z * [new tag] v1.12.0-rc6 -> v1.12.0-rc6 2025-09-07T07:36:04.3948131Z * [new tag] v1.12.0-rc7 -> v1.12.0-rc7 2025-09-07T07:36:04.3948806Z * [new tag] v1.12.0-rc8 -> v1.12.0-rc8 2025-09-07T07:36:04.3949618Z * [new tag] v1.12.1 -> v1.12.1 2025-09-07T07:36:04.3950635Z * [new tag] v1.12.1-rc1 -> v1.12.1-rc1 2025-09-07T07:36:04.3951640Z * [new tag] v1.12.1-rc2 -> v1.12.1-rc2 2025-09-07T07:36:04.3952700Z * [new tag] v1.12.1-rc3 -> v1.12.1-rc3 2025-09-07T07:36:04.3953752Z * [new tag] v1.12.1-rc4 -> v1.12.1-rc4 2025-09-07T07:36:04.3954383Z * [new tag] v1.12.1-rc5 -> v1.12.1-rc5 2025-09-07T07:36:04.3955441Z * [new tag] v1.13.0 -> v1.13.0 2025-09-07T07:36:04.3956353Z * [new tag] v1.13.0-rc1 -> v1.13.0-rc1 2025-09-07T07:36:04.3957232Z * [new tag] v1.13.0-rc2 -> v1.13.0-rc2 2025-09-07T07:36:04.3958165Z * [new tag] v1.13.0-rc3 -> v1.13.0-rc3 2025-09-07T07:36:04.3959195Z * [new tag] v1.13.0-rc4 -> v1.13.0-rc4 2025-09-07T07:36:04.3959946Z * [new tag] v1.13.0-rc5 -> v1.13.0-rc5 2025-09-07T07:36:04.3960806Z * [new tag] v1.13.0-rc6 -> v1.13.0-rc6 2025-09-07T07:36:04.3961781Z * [new tag] v1.13.1 -> v1.13.1 2025-09-07T07:36:04.3962561Z * [new tag] v1.13.1-rc1 -> v1.13.1-rc1 2025-09-07T07:36:04.3963462Z * [new tag] v1.2.0 -> v1.2.0 2025-09-07T07:36:04.3964506Z * [new tag] v1.2.0a0 -> v1.2.0a0 2025-09-07T07:36:04.3965585Z * [new tag] v1.3.0 -> v1.3.0 2025-09-07T07:36:04.3966578Z * [new tag] v1.3.0a0 -> v1.3.0a0 2025-09-07T07:36:04.3967354Z * [new tag] v1.3.1 -> v1.3.1 2025-09-07T07:36:04.3968297Z * [new tag] v1.4.0 -> v1.4.0 2025-09-07T07:36:04.3969215Z * [new tag] v1.4.0a0 -> v1.4.0a0 2025-09-07T07:36:04.3969973Z * [new tag] v1.4.1 -> v1.4.1 2025-09-07T07:36:04.3970956Z * [new tag] v1.5.0 -> v1.5.0 2025-09-07T07:36:04.3971904Z * [new tag] v1.5.0-rc1 -> v1.5.0-rc1 2025-09-07T07:36:04.3972918Z * [new tag] v1.5.0-rc2 -> v1.5.0-rc2 2025-09-07T07:36:04.3973913Z * [new tag] v1.5.0-rc3 -> v1.5.0-rc3 2025-09-07T07:36:04.3974780Z * [new tag] v1.5.0-rc4 -> v1.5.0-rc4 2025-09-07T07:36:04.3975542Z * [new tag] v1.5.0-rc5 -> v1.5.0-rc5 2025-09-07T07:36:04.3976513Z * [new tag] v1.5.1 -> v1.5.1 2025-09-07T07:36:04.3977287Z * [new tag] v1.5.1-rc1 -> v1.5.1-rc1 2025-09-07T07:36:04.3978115Z * [new tag] v1.6.0 -> v1.6.0 2025-09-07T07:36:04.3979159Z * [new tag] v1.6.0-rc1 -> v1.6.0-rc1 2025-09-07T07:36:04.3980109Z * [new tag] v1.6.0-rc2 -> v1.6.0-rc2 2025-09-07T07:36:04.3981064Z * [new tag] v1.6.0-rc3 -> v1.6.0-rc3 2025-09-07T07:36:04.3982084Z * [new tag] v1.6.0-rc4 -> v1.6.0-rc4 2025-09-07T07:36:04.3982962Z * [new tag] v1.6.0-rc5 -> v1.6.0-rc5 2025-09-07T07:36:04.3983904Z * [new tag] v1.6.0-rc6 -> v1.6.0-rc6 2025-09-07T07:36:04.3984701Z * [new tag] v1.6.0-rc7 -> v1.6.0-rc7 2025-09-07T07:36:04.3985802Z * [new tag] v1.7.0 -> v1.7.0 2025-09-07T07:36:04.3986808Z * [new tag] v1.7.0-rc1 -> v1.7.0-rc1 2025-09-07T07:36:04.3987936Z * [new tag] v1.7.0-rc2 -> v1.7.0-rc2 2025-09-07T07:36:04.3989411Z * [new tag] v1.7.0-rc3 -> v1.7.0-rc3 2025-09-07T07:36:04.3990205Z * [new tag] v1.7.0-rc4 -> v1.7.0-rc4 2025-09-07T07:36:04.3991232Z * [new tag] v1.7.1 -> v1.7.1 2025-09-07T07:36:04.3992421Z * [new tag] v1.7.1-rc1 -> v1.7.1-rc1 2025-09-07T07:36:04.3993309Z * [new tag] v1.7.1-rc2 -> v1.7.1-rc2 2025-09-07T07:36:04.3994063Z * [new tag] v1.7.1-rc3 -> v1.7.1-rc3 2025-09-07T07:36:04.3995064Z * [new tag] v1.8.0 -> v1.8.0 2025-09-07T07:36:04.3995858Z * [new tag] v1.8.0-rc1 -> v1.8.0-rc1 2025-09-07T07:36:04.3996861Z * [new tag] v1.8.0-rc2 -> v1.8.0-rc2 2025-09-07T07:36:04.3997786Z * [new tag] v1.8.0-rc3 -> v1.8.0-rc3 2025-09-07T07:36:04.3998665Z * [new tag] v1.8.0-rc4 -> v1.8.0-rc4 2025-09-07T07:36:04.3999467Z * [new tag] v1.8.0-rc5 -> v1.8.0-rc5 2025-09-07T07:36:04.4000127Z * [new tag] v1.8.1 -> v1.8.1 2025-09-07T07:36:04.4001148Z * [new tag] v1.8.1-rc1 -> v1.8.1-rc1 2025-09-07T07:36:04.4001943Z * [new tag] v1.8.1-rc2 -> v1.8.1-rc2 2025-09-07T07:36:04.4002610Z * [new tag] v1.8.1-rc3 -> v1.8.1-rc3 2025-09-07T07:36:04.4007542Z * [new tag] v1.8.2 -> v1.8.2 2025-09-07T07:36:04.4008327Z * [new tag] v1.8.2-rc1 -> v1.8.2-rc1 2025-09-07T07:36:04.4009364Z * [new tag] v1.9.0 -> v1.9.0 2025-09-07T07:36:04.4010358Z * [new tag] v1.9.0-rc1 -> v1.9.0-rc1 2025-09-07T07:36:04.4011378Z * [new tag] v1.9.0-rc2 -> v1.9.0-rc2 2025-09-07T07:36:04.4012353Z * [new tag] v1.9.0-rc3 -> v1.9.0-rc3 2025-09-07T07:36:04.4013084Z * [new tag] v1.9.0-rc4 -> v1.9.0-rc4 2025-09-07T07:36:04.4014113Z * [new tag] v1.9.1 -> v1.9.1 2025-09-07T07:36:04.4015464Z * [new tag] v1.9.1-rc1 -> v1.9.1-rc1 2025-09-07T07:36:04.4016233Z * [new tag] v1.9.1-rc2 -> v1.9.1-rc2 2025-09-07T07:36:04.4017213Z * [new tag] v2.0.0 -> v2.0.0 2025-09-07T07:36:04.4018323Z * [new tag] v2.0.0-rc1 -> v2.0.0-rc1 2025-09-07T07:36:04.4019342Z * [new tag] v2.0.0-rc2 -> v2.0.0-rc2 2025-09-07T07:36:04.4020323Z * [new tag] v2.0.0-rc3 -> v2.0.0-rc3 2025-09-07T07:36:04.4021234Z * [new tag] v2.0.0-rc4 -> v2.0.0-rc4 2025-09-07T07:36:04.4022264Z * [new tag] v2.0.0-rc5 -> v2.0.0-rc5 2025-09-07T07:36:04.4023023Z * [new tag] v2.0.0-rc6 -> v2.0.0-rc6 2025-09-07T07:36:04.4024055Z * [new tag] v2.0.1 -> v2.0.1 2025-09-07T07:36:04.4025119Z * [new tag] v2.0.1-rc1 -> v2.0.1-rc1 2025-09-07T07:36:04.4025903Z * [new tag] v2.0.1-rc2 -> v2.0.1-rc2 2025-09-07T07:36:04.4026798Z * [new tag] v2.0.1-rc3 -> v2.0.1-rc3 2025-09-07T07:36:04.4027554Z * [new tag] v2.0.1-rc4 -> v2.0.1-rc4 2025-09-07T07:36:04.4028904Z * [new tag] v2.1.0 -> v2.1.0 2025-09-07T07:36:04.4029827Z * [new tag] v2.1.0-rc1 -> v2.1.0-rc1 2025-09-07T07:36:04.4030773Z * [new tag] v2.1.0-rc2 -> v2.1.0-rc2 2025-09-07T07:36:04.4031810Z * [new tag] v2.1.0-rc3 -> v2.1.0-rc3 2025-09-07T07:36:04.4032807Z * [new tag] v2.1.0-rc4 -> v2.1.0-rc4 2025-09-07T07:36:04.4033762Z * [new tag] v2.1.0-rc5 -> v2.1.0-rc5 2025-09-07T07:36:04.4034634Z * [new tag] v2.1.0-rc6 -> v2.1.0-rc6 2025-09-07T07:36:04.4035526Z * [new tag] v2.1.1 -> v2.1.1 2025-09-07T07:36:04.4036524Z * [new tag] v2.1.1-rc1 -> v2.1.1-rc1 2025-09-07T07:36:04.4037392Z * [new tag] v2.1.1-rc2 -> v2.1.1-rc2 2025-09-07T07:36:04.4038464Z * [new tag] v2.1.1-rc3 -> v2.1.1-rc3 2025-09-07T07:36:04.4039427Z * [new tag] v2.1.1-rc4 -> v2.1.1-rc4 2025-09-07T07:36:04.4040330Z * [new tag] v2.1.1-rc5 -> v2.1.1-rc5 2025-09-07T07:36:04.4041091Z * [new tag] v2.1.1-rc6 -> v2.1.1-rc6 2025-09-07T07:36:04.4041999Z * [new tag] v2.1.2 -> v2.1.2 2025-09-07T07:36:04.4043008Z * [new tag] v2.1.2-rc1 -> v2.1.2-rc1 2025-09-07T07:36:04.4044030Z * [new tag] v2.1.2-rc2 -> v2.1.2-rc2 2025-09-07T07:36:04.4044798Z * [new tag] v2.1.2-rc3 -> v2.1.2-rc3 2025-09-07T07:36:04.4045772Z * [new tag] v2.2.0 -> v2.2.0 2025-09-07T07:36:04.4046744Z * [new tag] v2.2.0-rc1 -> v2.2.0-rc1 2025-09-07T07:36:04.4047634Z * [new tag] v2.2.0-rc2 -> v2.2.0-rc2 2025-09-07T07:36:04.4048548Z * [new tag] v2.2.0-rc3 -> v2.2.0-rc3 2025-09-07T07:36:04.4049430Z * [new tag] v2.2.0-rc4 -> v2.2.0-rc4 2025-09-07T07:36:04.4050363Z * [new tag] v2.2.0-rc5 -> v2.2.0-rc5 2025-09-07T07:36:04.4051258Z * [new tag] v2.2.0-rc6 -> v2.2.0-rc6 2025-09-07T07:36:04.4052031Z * [new tag] v2.2.0-rc7 -> v2.2.0-rc7 2025-09-07T07:36:04.4053239Z * [new tag] v2.2.0-rc8 -> v2.2.0-rc8 2025-09-07T07:36:04.4054282Z * [new tag] v2.2.1 -> v2.2.1 2025-09-07T07:36:04.4055363Z * [new tag] v2.2.1-rc1 -> v2.2.1-rc1 2025-09-07T07:36:04.4056130Z * [new tag] v2.2.1-rc2 -> v2.2.1-rc2 2025-09-07T07:36:04.4056855Z * [new tag] v2.2.1-rc3 -> v2.2.1-rc3 2025-09-07T07:36:04.4057676Z * [new tag] v2.2.2 -> v2.2.2 2025-09-07T07:36:04.4058742Z * [new tag] v2.2.2-rc1 -> v2.2.2-rc1 2025-09-07T07:36:04.4059531Z * [new tag] v2.2.2-rc2 -> v2.2.2-rc2 2025-09-07T07:36:04.4060331Z * [new tag] v2.2.2-rc3 -> v2.2.2-rc3 2025-09-07T07:36:04.4061208Z * [new tag] v2.3.0 -> v2.3.0 2025-09-07T07:36:04.4062247Z * [new tag] v2.3.0-rc1 -> v2.3.0-rc1 2025-09-07T07:36:04.4063286Z * [new tag] v2.3.0-rc10 -> v2.3.0-rc10 2025-09-07T07:36:04.4064280Z * [new tag] v2.3.0-rc11 -> v2.3.0-rc11 2025-09-07T07:36:04.4065027Z * [new tag] v2.3.0-rc12 -> v2.3.0-rc12 2025-09-07T07:36:04.4065995Z * [new tag] v2.3.0-rc2 -> v2.3.0-rc2 2025-09-07T07:36:04.4066941Z * [new tag] v2.3.0-rc3 -> v2.3.0-rc3 2025-09-07T07:36:04.4067883Z * [new tag] v2.3.0-rc4 -> v2.3.0-rc4 2025-09-07T07:36:04.4068774Z * [new tag] v2.3.0-rc5 -> v2.3.0-rc5 2025-09-07T07:36:04.4069527Z * [new tag] v2.3.0-rc6 -> v2.3.0-rc6 2025-09-07T07:36:04.4070480Z * [new tag] v2.3.0-rc7 -> v2.3.0-rc7 2025-09-07T07:36:04.4071408Z * [new tag] v2.3.0-rc8 -> v2.3.0-rc8 2025-09-07T07:36:04.4072159Z * [new tag] v2.3.0-rc9 -> v2.3.0-rc9 2025-09-07T07:36:04.4073029Z * [new tag] v2.3.1 -> v2.3.1 2025-09-07T07:36:04.4073864Z * [new tag] v2.3.1-rc1 -> v2.3.1-rc1 2025-09-07T07:36:04.4074780Z * [new tag] v2.3.1-rc2 -> v2.3.1-rc2 2025-09-07T07:36:04.4075740Z * [new tag] v2.3.1-rc3 -> v2.3.1-rc3 2025-09-07T07:36:04.4076679Z * [new tag] v2.4.0 -> v2.4.0 2025-09-07T07:36:04.4077630Z * [new tag] v2.4.0-rc1 -> v2.4.0-rc1 2025-09-07T07:36:04.4078556Z * [new tag] v2.4.0-rc2 -> v2.4.0-rc2 2025-09-07T07:36:04.4079472Z * [new tag] v2.4.0-rc3 -> v2.4.0-rc3 2025-09-07T07:36:04.4080355Z * [new tag] v2.4.0-rc4 -> v2.4.0-rc4 2025-09-07T07:36:04.4081365Z * [new tag] v2.4.0-rc5 -> v2.4.0-rc5 2025-09-07T07:36:04.4082286Z * [new tag] v2.4.0-rc6 -> v2.4.0-rc6 2025-09-07T07:36:04.4083260Z * [new tag] v2.4.0-rc7 -> v2.4.0-rc7 2025-09-07T07:36:04.4084263Z * [new tag] v2.4.0-rc8 -> v2.4.0-rc8 2025-09-07T07:36:04.4085225Z * [new tag] v2.4.0-rc9 -> v2.4.0-rc9 2025-09-07T07:36:04.4085994Z * [new tag] v2.4.1 -> v2.4.1 2025-09-07T07:36:04.4087020Z * [new tag] v2.4.1-rc1 -> v2.4.1-rc1 2025-09-07T07:36:04.4087966Z * [new tag] v2.4.1-rc2 -> v2.4.1-rc2 2025-09-07T07:36:04.4089008Z * [new tag] v2.4.1-rc3 -> v2.4.1-rc3 2025-09-07T07:36:04.4089931Z * [new tag] v2.5.0 -> v2.5.0 2025-09-07T07:36:04.4090844Z * [new tag] v2.5.0-rc1 -> v2.5.0-rc1 2025-09-07T07:36:04.4091599Z * [new tag] v2.5.0-rc10 -> v2.5.0-rc10 2025-09-07T07:36:04.4092556Z * [new tag] v2.5.0-rc2 -> v2.5.0-rc2 2025-09-07T07:36:04.4093458Z * [new tag] v2.5.0-rc3 -> v2.5.0-rc3 2025-09-07T07:36:04.4094384Z * [new tag] v2.5.0-rc4 -> v2.5.0-rc4 2025-09-07T07:36:04.4095321Z * [new tag] v2.5.0-rc5 -> v2.5.0-rc5 2025-09-07T07:36:04.4096308Z * [new tag] v2.5.0-rc6 -> v2.5.0-rc6 2025-09-07T07:36:04.4097279Z * [new tag] v2.5.0-rc7 -> v2.5.0-rc7 2025-09-07T07:36:04.4098327Z * [new tag] v2.5.0-rc8 -> v2.5.0-rc8 2025-09-07T07:36:04.4099294Z * [new tag] v2.5.0-rc9 -> v2.5.0-rc9 2025-09-07T07:36:04.4100044Z * [new tag] v2.5.1 -> v2.5.1 2025-09-07T07:36:04.4100807Z * [new tag] v2.5.1-rc1 -> v2.5.1-rc1 2025-09-07T07:36:04.4101554Z * [new tag] v2.6.0 -> v2.6.0 2025-09-07T07:36:04.4102572Z * [new tag] v2.6.0-rc1 -> v2.6.0-rc1 2025-09-07T07:36:04.4103845Z * [new tag] v2.6.0-rc2 -> v2.6.0-rc2 2025-09-07T07:36:04.4104841Z * [new tag] v2.6.0-rc3 -> v2.6.0-rc3 2025-09-07T07:36:04.4105743Z * [new tag] v2.6.0-rc4 -> v2.6.0-rc4 2025-09-07T07:36:04.4106976Z * [new tag] v2.6.0-rc5 -> v2.6.0-rc5 2025-09-07T07:36:04.4108022Z * [new tag] v2.6.0-rc6 -> v2.6.0-rc6 2025-09-07T07:36:04.4109108Z * [new tag] v2.6.0-rc7 -> v2.6.0-rc7 2025-09-07T07:36:04.4110132Z * [new tag] v2.6.0-rc8 -> v2.6.0-rc8 2025-09-07T07:36:04.4111111Z * [new tag] v2.6.0-rc9 -> v2.6.0-rc9 2025-09-07T07:36:04.4112246Z * [new tag] v2.7.0 -> v2.7.0 2025-09-07T07:36:04.4113496Z * [new tag] v2.7.0-rc1 -> v2.7.0-rc1 2025-09-07T07:36:04.4114287Z * [new tag] v2.7.0-rc10 -> v2.7.0-rc10 2025-09-07T07:36:04.4115363Z * [new tag] v2.7.0-rc2 -> v2.7.0-rc2 2025-09-07T07:36:04.4116361Z * [new tag] v2.7.0-rc3 -> v2.7.0-rc3 2025-09-07T07:36:04.4117334Z * [new tag] v2.7.0-rc4 -> v2.7.0-rc4 2025-09-07T07:36:04.4118291Z * [new tag] v2.7.0-rc5 -> v2.7.0-rc5 2025-09-07T07:36:04.4119236Z * [new tag] v2.7.0-rc6 -> v2.7.0-rc6 2025-09-07T07:36:04.4120176Z * [new tag] v2.7.0-rc7 -> v2.7.0-rc7 2025-09-07T07:36:04.4121168Z * [new tag] v2.7.0-rc8 -> v2.7.0-rc8 2025-09-07T07:36:04.4122151Z * [new tag] v2.7.0-rc9 -> v2.7.0-rc9 2025-09-07T07:36:04.4122920Z * [new tag] v2.7.1 -> v2.7.1 2025-09-07T07:36:04.4124002Z * [new tag] v2.7.1-rc1 -> v2.7.1-rc1 2025-09-07T07:36:04.4124999Z * [new tag] v2.7.1-rc2 -> v2.7.1-rc2 2025-09-07T07:36:04.4125993Z * [new tag] v2.7.1-rc3 -> v2.7.1-rc3 2025-09-07T07:36:04.4127041Z * [new tag] v2.7.1-rc4 -> v2.7.1-rc4 2025-09-07T07:36:04.4128026Z * [new tag] v2.7.1-rc5 -> v2.7.1-rc5 2025-09-07T07:36:04.4128861Z * [new tag] v2.8.0 -> v2.8.0 2025-09-07T07:36:04.4129838Z * [new tag] v2.8.0-rc1 -> v2.8.0-rc1 2025-09-07T07:36:04.4130855Z * [new tag] v2.8.0-rc2 -> v2.8.0-rc2 2025-09-07T07:36:04.4131905Z * [new tag] v2.8.0-rc3 -> v2.8.0-rc3 2025-09-07T07:36:04.4132876Z * [new tag] v2.8.0-rc4 -> v2.8.0-rc4 2025-09-07T07:36:04.4133882Z * [new tag] v2.8.0-rc5 -> v2.8.0-rc5 2025-09-07T07:36:04.4134876Z * [new tag] v2.8.0-rc6 -> v2.8.0-rc6 2025-09-07T07:36:04.4135842Z * [new tag] v2.8.0-rc7 -> v2.8.0-rc7 2025-09-07T07:36:04.4136819Z * [new tag] v2.8.0-rc8 -> v2.8.0-rc8 2025-09-07T07:36:04.4137910Z * [new tag] whc_flight_1 -> whc_flight_1 2025-09-07T07:36:04.4138936Z * [new tag] whc_flight_2 -> whc_flight_2 2025-09-07T07:36:04.4139818Z * [new tag] whc_flight_4 -> whc_flight_4 2025-09-07T07:36:04.4965532Z [command]/usr/bin/git rev-parse --verify --quiet 93fb23d6fae7c4e82c4239a1033e522088742634^{object} 2025-09-07T07:36:04.4996750Z 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:36:04.5001076Z ##[endgroup] 2025-09-07T07:36:04.5001272Z ##[group]Determining the checkout info 2025-09-07T07:36:04.5002299Z ##[endgroup] 2025-09-07T07:36:04.5006655Z [command]/usr/bin/git sparse-checkout disable 2025-09-07T07:36:04.5049614Z [command]/usr/bin/git config --local --unset-all extensions.worktreeConfig 2025-09-07T07:36:04.5080758Z ##[group]Checking out the ref 2025-09-07T07:36:04.5083906Z [command]/usr/bin/git checkout --progress --force 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:36:05.5372519Z Updating files: 85% (16573/19405) 2025-09-07T07:36:05.5501129Z Updating files: 86% (16689/19405) 2025-09-07T07:36:05.5632355Z Updating files: 87% (16883/19405) 2025-09-07T07:36:05.5734103Z Updating files: 88% (17077/19405) 2025-09-07T07:36:05.5863971Z Updating files: 89% (17271/19405) 2025-09-07T07:36:05.6021489Z Updating files: 90% (17465/19405) 2025-09-07T07:36:05.6132104Z Updating files: 91% (17659/19405) 2025-09-07T07:36:05.6263844Z Updating files: 92% (17853/19405) 2025-09-07T07:36:05.6434058Z Updating files: 93% (18047/19405) 2025-09-07T07:36:05.6616232Z Updating files: 94% (18241/19405) 2025-09-07T07:36:05.6759065Z Updating files: 95% (18435/19405) 2025-09-07T07:36:05.6907061Z Updating files: 96% (18629/19405) 2025-09-07T07:36:05.7068048Z Updating files: 97% (18823/19405) 2025-09-07T07:36:05.7294664Z Updating files: 98% (19017/19405) 2025-09-07T07:36:05.7440003Z Updating files: 99% (19211/19405) 2025-09-07T07:36:05.7440455Z Updating files: 100% (19405/19405) 2025-09-07T07:36:05.7440699Z Updating files: 100% (19405/19405), done. 2025-09-07T07:36:05.7645953Z Note: switching to '93fb23d6fae7c4e82c4239a1033e522088742634'. 2025-09-07T07:36:05.7646191Z 2025-09-07T07:36:05.7646337Z You are in 'detached HEAD' state. You can look around, make experimental 2025-09-07T07:36:05.7646693Z changes and commit them, and you can discard any commits you make in this 2025-09-07T07:36:05.7647030Z state without impacting any branches by switching back to a branch. 2025-09-07T07:36:05.7647232Z 2025-09-07T07:36:05.7647367Z If you want to create a new branch to retain commits you create, you may 2025-09-07T07:36:05.7647674Z do so (now or later) by using -c with the switch command. Example: 2025-09-07T07:36:05.7647855Z 2025-09-07T07:36:05.7647937Z git switch -c 2025-09-07T07:36:05.7648066Z 2025-09-07T07:36:05.7648136Z Or undo this operation with: 2025-09-07T07:36:05.7648250Z 2025-09-07T07:36:05.7648318Z git switch - 2025-09-07T07:36:05.7648405Z 2025-09-07T07:36:05.7648561Z Turn off this advice by setting config variable advice.detachedHead to false 2025-09-07T07:36:05.7648767Z 2025-09-07T07:36:05.7648883Z HEAD is now at 93fb23d6fae Build vLLM nightly wheels (#162000) 2025-09-07T07:36:05.7756004Z ##[endgroup] 2025-09-07T07:36:05.7756303Z ##[group]Setting up auth for fetching submodules 2025-09-07T07:36:05.7761106Z [command]/usr/bin/git config --global http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-09-07T07:36:05.7808571Z [command]/usr/bin/git config --global --unset-all url.https://github.com/.insteadOf 2025-09-07T07:36:05.7839571Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf git@github.com: 2025-09-07T07:36:05.7871310Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf org-21003710@github.com: 2025-09-07T07:36:05.7900850Z ##[endgroup] 2025-09-07T07:36:05.7901107Z ##[group]Fetching submodules 2025-09-07T07:36:05.7903865Z [command]/usr/bin/git submodule sync --recursive 2025-09-07T07:36:05.8275708Z [command]/usr/bin/git -c protocol.version=2 submodule update --init --force --recursive 2025-09-07T07:36:05.8633887Z Submodule 'android/libs/fbjni' (https://github.com/facebookincubator/fbjni.git) registered for path 'android/libs/fbjni' 2025-09-07T07:36:05.8635784Z Submodule 'third_party/NNPACK_deps/FP16' (https://github.com/Maratyszcza/FP16.git) registered for path 'third_party/FP16' 2025-09-07T07:36:05.8639096Z Submodule 'third_party/NNPACK_deps/FXdiv' (https://github.com/Maratyszcza/FXdiv.git) registered for path 'third_party/FXdiv' 2025-09-07T07:36:05.8948289Z Submodule 'third_party/NNPACK' (https://github.com/Maratyszcza/NNPACK.git) registered for path 'third_party/NNPACK' 2025-09-07T07:36:05.8950709Z Submodule 'third_party/NVTX' (https://github.com/NVIDIA/NVTX.git) registered for path 'third_party/NVTX' 2025-09-07T07:36:05.8954371Z Submodule 'third_party/VulkanMemoryAllocator' (https://github.com/GPUOpen-LibrariesAndSDKs/VulkanMemoryAllocator.git) registered for path 'third_party/VulkanMemoryAllocator' 2025-09-07T07:36:05.8957921Z Submodule 'third_party/XNNPACK' (https://github.com/google/XNNPACK.git) registered for path 'third_party/XNNPACK' 2025-09-07T07:36:05.8961564Z Submodule 'third_party/aiter' (https://github.com/ROCm/aiter.git) registered for path 'third_party/aiter' 2025-09-07T07:36:05.8965332Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/benchmark' 2025-09-07T07:36:05.8976452Z Submodule 'third_party/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/composable_kernel' 2025-09-07T07:36:05.8978641Z Submodule 'third_party/cpp-httplib' (https://github.com/yhirose/cpp-httplib.git) registered for path 'third_party/cpp-httplib' 2025-09-07T07:36:05.8982969Z Submodule 'third_party/cpuinfo' (https://github.com/pytorch/cpuinfo.git) registered for path 'third_party/cpuinfo' 2025-09-07T07:36:05.8987008Z Submodule 'third_party/cudnn_frontend' (https://github.com/NVIDIA/cudnn-frontend.git) registered for path 'third_party/cudnn_frontend' 2025-09-07T07:36:05.8991135Z Submodule 'third_party/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/cutlass' 2025-09-07T07:36:05.8995139Z Submodule 'third_party/fbgemm' (https://github.com/pytorch/fbgemm) registered for path 'third_party/fbgemm' 2025-09-07T07:36:05.8999350Z Submodule 'third_party/flash-attention' (https://github.com/Dao-AILab/flash-attention.git) registered for path 'third_party/flash-attention' 2025-09-07T07:36:05.9008480Z Submodule 'third_party/flatbuffers' (https://github.com/google/flatbuffers.git) registered for path 'third_party/flatbuffers' 2025-09-07T07:36:05.9012617Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/fmt' 2025-09-07T07:36:05.9017095Z Submodule 'third_party/gemmlowp/gemmlowp' (https://github.com/google/gemmlowp.git) registered for path 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:36:05.9024064Z Submodule 'third_party/gloo' (https://github.com/pytorch/gloo) registered for path 'third_party/gloo' 2025-09-07T07:36:05.9028580Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/googletest' 2025-09-07T07:36:05.9032967Z Submodule 'third_party/ideep' (https://github.com/intel/ideep) registered for path 'third_party/ideep' 2025-09-07T07:36:05.9043361Z Submodule 'third_party/ittapi' (https://github.com/intel/ittapi.git) registered for path 'third_party/ittapi' 2025-09-07T07:36:05.9048201Z Submodule 'third_party/kineto' (https://github.com/pytorch/kineto) registered for path 'third_party/kineto' 2025-09-07T07:36:05.9052699Z Submodule 'third_party/kleidiai' (https://github.com/ARM-software/kleidiai.git) registered for path 'third_party/kleidiai' 2025-09-07T07:36:05.9057358Z Submodule 'third_party/mimalloc' (https://github.com/microsoft/mimalloc.git) registered for path 'third_party/mimalloc' 2025-09-07T07:36:05.9062216Z Submodule 'third_party/nlohmann' (https://github.com/nlohmann/json.git) registered for path 'third_party/nlohmann' 2025-09-07T07:36:05.9066951Z Submodule 'third_party/onnx' (https://github.com/onnx/onnx.git) registered for path 'third_party/onnx' 2025-09-07T07:36:05.9071961Z Submodule 'third_party/opentelemetry-cpp' (https://github.com/open-telemetry/opentelemetry-cpp.git) registered for path 'third_party/opentelemetry-cpp' 2025-09-07T07:36:05.9081854Z Submodule 'third_party/pocketfft' (https://github.com/mreineck/pocketfft) registered for path 'third_party/pocketfft' 2025-09-07T07:36:05.9087208Z Submodule 'third_party/protobuf' (https://github.com/protocolbuffers/protobuf.git) registered for path 'third_party/protobuf' 2025-09-07T07:36:05.9092380Z Submodule 'third_party/NNPACK_deps/psimd' (https://github.com/Maratyszcza/psimd.git) registered for path 'third_party/psimd' 2025-09-07T07:36:05.9097673Z Submodule 'third_party/NNPACK_deps/pthreadpool' (https://github.com/Maratyszcza/pthreadpool.git) registered for path 'third_party/pthreadpool' 2025-09-07T07:36:05.9102818Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/pybind11' 2025-09-07T07:36:05.9108295Z Submodule 'third_party/python-peachpy' (https://github.com/malfet/PeachPy.git) registered for path 'third_party/python-peachpy' 2025-09-07T07:36:05.9118866Z Submodule 'third_party/sleef' (https://github.com/shibatch/sleef) registered for path 'third_party/sleef' 2025-09-07T07:36:05.9127003Z Submodule 'third_party/tensorpipe' (https://github.com/pytorch/tensorpipe.git) registered for path 'third_party/tensorpipe' 2025-09-07T07:36:05.9168010Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/android/libs/fbjni'... 2025-09-07T07:36:06.1212128Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FXdiv'... 2025-09-07T07:36:06.1212624Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FP16'... 2025-09-07T07:36:06.1213067Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/psimd'... 2025-09-07T07:36:06.1213477Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NNPACK'... 2025-09-07T07:36:06.1471927Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NVTX'... 2025-09-07T07:36:06.2587043Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pthreadpool'... 2025-09-07T07:36:06.2587618Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pocketfft'... 2025-09-07T07:36:06.2588084Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/python-peachpy'... 2025-09-07T07:36:06.2588522Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep'... 2025-09-07T07:36:06.2588950Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gemmlowp/gemmlowp'... 2025-09-07T07:36:06.2589386Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gloo'... 2025-09-07T07:36:06.2673228Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/VulkanMemoryAllocator'... 2025-09-07T07:36:07.0376271Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/benchmark'... 2025-09-07T07:36:07.0376790Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ittapi'... 2025-09-07T07:36:07.0377532Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe'... 2025-09-07T07:36:07.0378034Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kleidiai'... 2025-09-07T07:36:07.0378482Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention'... 2025-09-07T07:36:07.0378910Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpuinfo'... 2025-09-07T07:36:07.0379336Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpp-httplib'... 2025-09-07T07:36:07.0379751Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/sleef'... 2025-09-07T07:36:07.0380153Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/googletest'... 2025-09-07T07:36:07.0380584Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pybind11'... 2025-09-07T07:36:07.0380995Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/mimalloc'... 2025-09-07T07:36:07.0381395Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fmt'... 2025-09-07T07:36:07.0381823Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cudnn_frontend'... 2025-09-07T07:36:07.1377468Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/XNNPACK'... 2025-09-07T07:36:14.8832692Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto'... 2025-09-07T07:36:14.8833213Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flatbuffers'... 2025-09-07T07:36:14.8833654Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm'... 2025-09-07T07:36:14.8834067Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cutlass'... 2025-09-07T07:36:14.8834478Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx'... 2025-09-07T07:36:14.8834948Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/composable_kernel'... 2025-09-07T07:36:14.8835843Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter'... 2025-09-07T07:36:14.8836280Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp'... 2025-09-07T07:36:14.8836722Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/nlohmann'... 2025-09-07T07:36:14.8837136Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf'... 2025-09-07T07:36:14.9014561Z Submodule path 'android/libs/fbjni': checked out '7e1e1fe3858c63c251c637ae41a20de425dde96f' 2025-09-07T07:36:14.9165231Z Submodule path 'third_party/FP16': checked out '4dfe081cf6bcd15db339cf2680b9281b8451eeb3' 2025-09-07T07:36:14.9275601Z Submodule path 'third_party/FXdiv': checked out 'b408327ac2a15ec3e43352421954f5b1967701d1' 2025-09-07T07:36:14.9549510Z Submodule path 'third_party/NNPACK': checked out 'c07e3a0400713d546e0dea2d5466dd22ea389c73' 2025-09-07T07:36:15.0299929Z Submodule path 'third_party/NVTX': checked out '2942f167cc30c5e3a44a2aecd5b0d9c07ff61a07' 2025-09-07T07:36:15.0795061Z Submodule path 'third_party/VulkanMemoryAllocator': checked out '1d8f600fd424278486eade7ed3e877c99f0846b1' 2025-09-07T07:36:15.8194568Z Submodule path 'third_party/XNNPACK': checked out '51a0103656eff6fc9bfd39a4597923c4b542c883' 2025-09-07T07:36:15.9680946Z Submodule path 'third_party/aiter': checked out '01aae101b9e5e94d6c16a9514c9fb8df99c93150' 2025-09-07T07:36:15.9707876Z Submodule '3rdparty/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:36:15.9744728Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter/3rdparty/composable_kernel'... 2025-09-07T07:36:19.0414484Z Submodule path 'third_party/aiter/3rdparty/composable_kernel': checked out 'cffe8fa2a442ac8e80dd236a1a5d24fe3d7e0cbf' 2025-09-07T07:36:19.0683816Z Submodule path 'third_party/benchmark': checked out '299e5928955cc62af9968370293b916f5130916f' 2025-09-07T07:36:19.3949749Z Submodule path 'third_party/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-09-07T07:36:19.4436552Z Submodule path 'third_party/cpp-httplib': checked out '89c932f313c6437c38f2982869beacc89c2f2246' 2025-09-07T07:36:19.5357502Z Submodule path 'third_party/cpuinfo': checked out '5e3d2445e6a84d9599bee2bf78edbb4d80865e1d' 2025-09-07T07:36:19.5805042Z Submodule path 'third_party/cudnn_frontend': checked out 'f937055efc6d414d11f4c6577e3977fe74f35fb6' 2025-09-07T07:36:20.1757349Z Submodule path 'third_party/cutlass': checked out 'e51efbfe18fe4f4cbb66ab814c55bf4aa0185491' 2025-09-07T07:36:20.3217280Z Submodule path 'third_party/fbgemm': checked out '4b39c551efe15e6bbade20565b0ceb2d8ce3352d' 2025-09-07T07:36:20.3246180Z Submodule 'external/asmjit' (https://github.com/asmjit/asmjit.git) registered for path 'third_party/fbgemm/external/asmjit' 2025-09-07T07:36:20.3247937Z Submodule 'external/composable_kernel' (https://github.com/jwfromm/composable_kernel.git) registered for path 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:36:20.3250660Z Submodule 'external/cpuinfo' (https://github.com/pytorch/cpuinfo) registered for path 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:36:20.3253589Z Submodule 'external/cutlass' (https://github.com/jwfromm/cutlass) registered for path 'third_party/fbgemm/external/cutlass' 2025-09-07T07:36:20.3256617Z Submodule 'external/googletest' (https://github.com/google/googletest) registered for path 'third_party/fbgemm/external/googletest' 2025-09-07T07:36:20.3259816Z Submodule 'external/hipify_torch' (https://github.com/ROCmSoftwarePlatform/hipify_torch.git) registered for path 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:36:20.3262721Z Submodule 'external/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/fbgemm/external/json' 2025-09-07T07:36:20.3303519Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/asmjit'... 2025-09-07T07:36:21.2744936Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/hipify_torch'... 2025-09-07T07:36:21.2745540Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cpuinfo'... 2025-09-07T07:36:21.2746050Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/googletest'... 2025-09-07T07:36:21.2746572Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/composable_kernel'... 2025-09-07T07:36:21.3744959Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cutlass'... 2025-09-07T07:36:21.8822275Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/json'... 2025-09-07T07:36:25.3485835Z Submodule path 'third_party/fbgemm/external/asmjit': checked out 'a3199e8857792cd10b7589ff5d58343d2c9008ea' 2025-09-07T07:36:25.6160304Z Submodule path 'third_party/fbgemm/external/composable_kernel': checked out 'b1281b8b08d973a7064f864f47eeb30f3e2596e9' 2025-09-07T07:36:25.7097536Z Submodule path 'third_party/fbgemm/external/cpuinfo': checked out '6543fec09b2f04ac4a666882998b534afc9c1349' 2025-09-07T07:36:26.3068161Z Submodule path 'third_party/fbgemm/external/cutlass': checked out '311f3c8e51dc0eb56310cfc6980bf63d0fbd7917' 2025-09-07T07:36:26.3524176Z Submodule path 'third_party/fbgemm/external/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-09-07T07:36:26.3665756Z Submodule path 'third_party/fbgemm/external/hipify_torch': checked out '63b6a7b541fa7f08f8475ca7d74054db36ff2691' 2025-09-07T07:36:26.4747191Z Submodule path 'third_party/fbgemm/external/json': checked out '9cca280a4d0ccf0c08f47a99aa71d1b0e52f8d03' 2025-09-07T07:36:26.5496732Z Submodule path 'third_party/flash-attention': checked out '979702c87a8713a8e0a5e9fee122b90d2ef13be5' 2025-09-07T07:36:26.5522556Z Submodule 'csrc/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:36:26.5523861Z Submodule 'csrc/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:36:26.5559574Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/composable_kernel'... 2025-09-07T07:36:29.2995217Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/cutlass'... 2025-09-07T07:36:29.5444896Z Submodule path 'third_party/flash-attention/csrc/composable_kernel': checked out '888317e698e9803c62bd38568abc9e05d7709f33' 2025-09-07T07:36:30.0802007Z Submodule path 'third_party/flash-attention/csrc/cutlass': checked out 'c506e16788cb08416a4a57e11a9067beeee29420' 2025-09-07T07:36:30.2220973Z Submodule path 'third_party/flatbuffers': checked out 'a2cd1ea3b6d3fee220106b5fed3f7ce8da9eb757' 2025-09-07T07:36:30.2554687Z Submodule path 'third_party/fmt': checked out '40626af88bd7df9a5fb80be7b25ac85b122d6c21' 2025-09-07T07:36:30.2948648Z Submodule path 'third_party/gemmlowp/gemmlowp': checked out '3fb5c176c17c765a3492cd2f0321b0dab712f350' 2025-09-07T07:36:30.3230126Z Submodule path 'third_party/gloo': checked out 'c7b7b022c124d9643957d9bd55f57ac59fce8fa2' 2025-09-07T07:36:30.3689425Z Submodule path 'third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-09-07T07:36:30.3842550Z Submodule path 'third_party/ideep': checked out '719d8e6cd7f7a0e01b155657526d693acf97c2b3' 2025-09-07T07:36:30.3865327Z Submodule 'mkl-dnn' (https://github.com/intel/mkl-dnn.git) registered for path 'third_party/ideep/mkl-dnn' 2025-09-07T07:36:30.3898386Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep/mkl-dnn'... 2025-09-07T07:36:40.7427807Z Submodule path 'third_party/ideep/mkl-dnn': checked out '8d263e693366ef8db40acc569cc7d8edf644556d' 2025-09-07T07:36:40.7661127Z Submodule path 'third_party/ittapi': checked out 'dec1d23ca65ab069d225dfe40dea14f455170959' 2025-09-07T07:36:40.8477653Z Submodule path 'third_party/kineto': checked out '5e7501833f1021ce6f618572d3baf657b6319658' 2025-09-07T07:36:40.8503566Z Submodule 'libkineto/third_party/dynolog' (https://github.com/facebookincubator/dynolog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:36:40.8505104Z Submodule 'libkineto/third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:36:40.8508382Z Submodule 'libkineto/third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:36:40.8546321Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog'... 2025-09-07T07:36:41.4281370Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/fmt'... 2025-09-07T07:36:41.6269147Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/googletest'... 2025-09-07T07:36:41.7043976Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog': checked out '7d04a0053a845370ae06ce317a22a48e9edcc74e' 2025-09-07T07:36:41.7067611Z Submodule 'third_party/DCGM' (https://github.com/NVIDIA/DCGM.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:36:41.7069523Z Submodule 'third_party/cpr' (https://github.com/libcpr/cpr.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:36:41.7072531Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:36:41.7075625Z Submodule 'third_party/gflags' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:36:41.7078812Z Submodule 'third_party/glog' (https://github.com/google/glog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:36:41.7082138Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:36:41.7085315Z Submodule 'third_party/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:36:41.7088682Z Submodule 'third_party/pfs' (https://github.com/dtrugman/pfs.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:36:41.7128817Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM'... 2025-09-07T07:36:42.7865748Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/pfs'... 2025-09-07T07:36:42.7866541Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags'... 2025-09-07T07:36:42.7867234Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/cpr'... 2025-09-07T07:36:42.7867920Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/glog'... 2025-09-07T07:36:42.7868580Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/googletest'... 2025-09-07T07:36:42.7869238Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/fmt'... 2025-09-07T07:36:42.8866448Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/json'... 2025-09-07T07:36:46.7563592Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM': checked out 'ffde4e54bc7249a6039a5e6b45b395141e1217f9' 2025-09-07T07:36:46.7775087Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr': checked out '871ed52d350214a034f6ef8a3b8f51c5ce1bd400' 2025-09-07T07:36:46.8139412Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt': checked out 'cd4af11efc9c622896a3e4cb599fa28668ca3d05' 2025-09-07T07:36:46.8296924Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags': checked out 'e171aa2d15ed9eb17054558e0b3a6a413bb01067' 2025-09-07T07:36:46.8317473Z Submodule 'doc' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:36:46.8353022Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc'... 2025-09-07T07:36:47.0897895Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc': checked out '8411df715cf522606e3b1aca386ddfc0b63d34b4' 2025-09-07T07:36:47.1119296Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog': checked out 'b33e3bad4c46c8a6345525fd822af355e5ef9446' 2025-09-07T07:36:47.1524585Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest': checked out '58d77fa8070e8cec2dc1ed015d66b454c8d78850' 2025-09-07T07:36:47.2524248Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json': checked out '4f8fba14066156b73f1189a2b8bd568bde5284c5' 2025-09-07T07:36:47.2717967Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs': checked out 'f68a2fa8ea36c783bdd760371411fcb495aa3150' 2025-09-07T07:36:47.3105772Z Submodule path 'third_party/kineto/libkineto/third_party/fmt': checked out '0041a40c1350ba702d475b9c4ad62da77caea164' 2025-09-07T07:36:47.3668530Z Submodule path 'third_party/kineto/libkineto/third_party/googletest': checked out '7aca84427f224eeed3144123d5230d5871e93347' 2025-09-07T07:36:47.4130983Z Submodule path 'third_party/kleidiai': checked out 'cca02c2f69dd18e1f12647c1c0bdc8cf90e680c7' 2025-09-07T07:36:47.4592520Z Submodule path 'third_party/mimalloc': checked out 'fbd8b99c2b828428947d70fdc046bb55609be93e' 2025-09-07T07:36:47.5694939Z Submodule path 'third_party/nlohmann': checked out '55f93686c01528224f448c19128836e7df245f72' 2025-09-07T07:36:48.0213807Z Submodule path 'third_party/onnx': checked out 'e709452ef2bbc1d113faf678c24e6d3467696e83' 2025-09-07T07:36:48.0255530Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/onnx/third_party/pybind11' 2025-09-07T07:36:48.0293250Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx/third_party/pybind11'... 2025-09-07T07:36:48.8117629Z Submodule path 'third_party/onnx/third_party/pybind11': checked out 'a2e59f0e7065404b44dfe92a28aca47ba1378dc4' 2025-09-07T07:36:48.8870876Z Submodule path 'third_party/opentelemetry-cpp': checked out 'a799f4aed9c94b765dcdaabaeab7d5e7e2310878' 2025-09-07T07:36:48.8897171Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark) registered for path 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:36:48.8898462Z Submodule 'third_party/googletest' (https://github.com/google/googletest) registered for path 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:36:48.8901314Z Submodule 'third_party/ms-gsl' (https://github.com/microsoft/GSL) registered for path 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:36:48.8904707Z Submodule 'third_party/nlohmann-json' (https://github.com/nlohmann/json) registered for path 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:36:48.8907728Z Submodule 'third_party/opentelemetry-proto' (https://github.com/open-telemetry/opentelemetry-proto) registered for path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:36:48.8921887Z Submodule 'third_party/opentracing-cpp' (https://github.com/opentracing/opentracing-cpp.git) registered for path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:36:48.8922786Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:36:48.8923520Z Submodule 'tools/vcpkg' (https://github.com/Microsoft/vcpkg) registered for path 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:36:48.8955629Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/benchmark'... 2025-09-07T07:36:49.6881334Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentracing-cpp'... 2025-09-07T07:36:49.6882052Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/ms-gsl'... 2025-09-07T07:36:49.6882685Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentelemetry-proto'... 2025-09-07T07:36:49.6883301Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp'... 2025-09-07T07:36:49.7881704Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/googletest'... 2025-09-07T07:36:49.9403866Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/nlohmann-json'... 2025-09-07T07:36:54.2464668Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/tools/vcpkg'... 2025-09-07T07:36:54.8804356Z Submodule path 'third_party/opentelemetry-cpp/third_party/benchmark': checked out 'd572f4777349d43653b21d6c2fc63020ab326db2' 2025-09-07T07:36:54.9200546Z Submodule path 'third_party/opentelemetry-cpp/third_party/googletest': checked out 'b796f7d44681514f58a683a3a71ff17c94edb0c1' 2025-09-07T07:36:54.9379207Z Submodule path 'third_party/opentelemetry-cpp/third_party/ms-gsl': checked out '6f4529395c5b7c2d661812257cd6780c67e54afa' 2025-09-07T07:36:55.0415020Z Submodule path 'third_party/opentelemetry-cpp/third_party/nlohmann-json': checked out 'bc889afb4c5bf1c0d8ee29ef35eaaf4c8bef8a5d' 2025-09-07T07:36:55.0581252Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto': checked out '4ca4f0335c63cda7ab31ea7ed70d6553aee14dce' 2025-09-07T07:36:55.0754353Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp': checked out '06b57f48ded1fa3bdd3d4346f6ef29e40e08eaf5' 2025-09-07T07:36:55.0939821Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp': checked out 'c9ffcdda9086ffd9e1283ea7a0276d831f3c8a8d' 2025-09-07T07:36:55.0963443Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:36:55.0964871Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:36:55.1002828Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-09-07T07:36:56.4030599Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest'... 2025-09-07T07:36:56.6266068Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'eefb26f82b233268fc98577d265352720d477ba4' 2025-09-07T07:36:56.6732316Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-09-07T07:36:57.2035844Z Submodule path 'third_party/opentelemetry-cpp/tools/vcpkg': checked out '8eb57355a4ffb410a2e94c07b4dca2dffbee8e50' 2025-09-07T07:36:57.2183705Z Submodule path 'third_party/pocketfft': checked out '0fa0ef591e38c2758e3184c6c23e497b9f732ffa' 2025-09-07T07:36:57.4789694Z Submodule path 'third_party/protobuf': checked out 'd1eca4e4b421cd2997495c4b4e65cea6be4e9b8a' 2025-09-07T07:36:57.4819483Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:36:57.4821130Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/protobuf/third_party/googletest' 2025-09-07T07:36:57.4858973Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/benchmark'... 2025-09-07T07:36:58.5353333Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/googletest'... 2025-09-07T07:36:58.5509954Z Submodule path 'third_party/protobuf/third_party/benchmark': checked out '5b7683f49e1e9223cf9927b24f6fd3d6bd82e3f8' 2025-09-07T07:36:58.6167108Z Submodule path 'third_party/protobuf/third_party/googletest': checked out '5ec7f0c4a113e2f18ac2c6cc7df51ad6afc24081' 2025-09-07T07:36:58.6289600Z Submodule path 'third_party/psimd': checked out '072586a71b55b7f8c584153d223e95687148a900' 2025-09-07T07:36:58.6438607Z Submodule path 'third_party/pthreadpool': checked out '4fe0e1e183925bf8cfa6aae24237e724a96479b8' 2025-09-07T07:36:58.6839615Z Submodule path 'third_party/pybind11': checked out 'f5fbe867d2d26e4a0a9177a51f6e568868ad3dc8' 2025-09-07T07:36:58.7136137Z Submodule path 'third_party/python-peachpy': checked out 'f45429b087dd7d5bc78bb40dc7cf06425c252d67' 2025-09-07T07:36:58.7564954Z Submodule path 'third_party/sleef': checked out '5a1d179df9cf652951b59010a2d2075372d67f68' 2025-09-07T07:36:58.7868018Z Submodule path 'third_party/tensorpipe': checked out 'af0118d13e52f5a08841464a768e01a0bf3e3075' 2025-09-07T07:36:58.7890283Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:36:58.7891528Z Submodule 'third_party/libnop' (https://github.com/google/libnop.git) registered for path 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:36:58.7894364Z Submodule 'third_party/libuv' (https://github.com/libuv/libuv.git) registered for path 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:36:58.7897672Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:36:58.7932928Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/googletest'... 2025-09-07T07:36:59.4952081Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libnop'... 2025-09-07T07:36:59.5218336Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libuv'... 2025-09-07T07:36:59.6915268Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11'... 2025-09-07T07:36:59.7459386Z Submodule path 'third_party/tensorpipe/third_party/googletest': checked out 'aee0f9d9b5b87796ee8a0ab26b7587ec30e8858e' 2025-09-07T07:36:59.7640405Z Submodule path 'third_party/tensorpipe/third_party/libnop': checked out '910b55815be16109f04f4180e9adee14fb4ce281' 2025-09-07T07:36:59.8334125Z Submodule path 'third_party/tensorpipe/third_party/libuv': checked out '5152db2cbfeb5582e9c27c5ea1dba2cd9e10759b' 2025-09-07T07:36:59.8643257Z Submodule path 'third_party/tensorpipe/third_party/pybind11': checked out 'a23996fce38ff6ccfbcdc09f1e63f2c4be5ea2ef' 2025-09-07T07:36:59.8664447Z Submodule 'tools/clang' (https://github.com/wjakob/clang-cindex-python3) registered for path 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:36:59.8698902Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11/tools/clang'... 2025-09-07T07:37:00.0686581Z Submodule path 'third_party/tensorpipe/third_party/pybind11/tools/clang': checked out '6a00cbc4a9b8e68b71caf7f774b3f9c753ae84d5' 2025-09-07T07:37:00.0733261Z [command]/usr/bin/git submodule foreach --recursive git config --local gc.auto 0 2025-09-07T07:37:00.1094195Z Entering 'android/libs/fbjni' 2025-09-07T07:37:00.1145162Z Entering 'third_party/FP16' 2025-09-07T07:37:00.1196696Z Entering 'third_party/FXdiv' 2025-09-07T07:37:00.1248267Z Entering 'third_party/NNPACK' 2025-09-07T07:37:00.1297906Z Entering 'third_party/NVTX' 2025-09-07T07:37:00.1348548Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:37:00.1395876Z Entering 'third_party/XNNPACK' 2025-09-07T07:37:00.1457731Z Entering 'third_party/aiter' 2025-09-07T07:37:00.1508124Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:37:00.1562868Z Entering 'third_party/benchmark' 2025-09-07T07:37:00.1611257Z Entering 'third_party/composable_kernel' 2025-09-07T07:37:00.1669906Z Entering 'third_party/cpp-httplib' 2025-09-07T07:37:00.1719171Z Entering 'third_party/cpuinfo' 2025-09-07T07:37:00.1770494Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:37:00.1820309Z Entering 'third_party/cutlass' 2025-09-07T07:37:00.1877527Z Entering 'third_party/fbgemm' 2025-09-07T07:37:00.1931461Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:37:00.1978490Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:37:00.2033680Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:37:00.2079745Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:37:00.2137574Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:37:00.2184251Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:37:00.2229958Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:37:00.2280955Z Entering 'third_party/flash-attention' 2025-09-07T07:37:00.2332320Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:37:00.2384352Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:37:00.2441290Z Entering 'third_party/flatbuffers' 2025-09-07T07:37:00.2493102Z Entering 'third_party/fmt' 2025-09-07T07:37:00.2543418Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:37:00.2593789Z Entering 'third_party/gloo' 2025-09-07T07:37:00.2645482Z Entering 'third_party/googletest' 2025-09-07T07:37:00.2695007Z Entering 'third_party/ideep' 2025-09-07T07:37:00.2745296Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:37:00.2800885Z Entering 'third_party/ittapi' 2025-09-07T07:37:00.2848997Z Entering 'third_party/kineto' 2025-09-07T07:37:00.2897925Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:37:00.2943598Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:37:00.2991350Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:37:00.3042750Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:37:00.3089124Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:37:00.3140053Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:37:00.3188776Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:37:00.3235141Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:37:00.3282485Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:37:00.3332477Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:37:00.3383025Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:37:00.3431716Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:37:00.3484630Z Entering 'third_party/kleidiai' 2025-09-07T07:37:00.3537925Z Entering 'third_party/mimalloc' 2025-09-07T07:37:00.3590571Z Entering 'third_party/nlohmann' 2025-09-07T07:37:00.3642413Z Entering 'third_party/onnx' 2025-09-07T07:37:00.3707101Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:37:00.3757002Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:37:00.3807684Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:37:00.3853320Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:37:00.3902861Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:37:00.3948546Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:37:00.3994930Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:37:00.4041535Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:37:00.4087005Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:37:00.4135042Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:37:00.4184735Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:37:00.4234688Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:37:00.4301152Z Entering 'third_party/pocketfft' 2025-09-07T07:37:00.4352898Z Entering 'third_party/protobuf' 2025-09-07T07:37:00.4407510Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:37:00.4453626Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:37:00.4508991Z Entering 'third_party/psimd' 2025-09-07T07:37:00.4559105Z Entering 'third_party/pthreadpool' 2025-09-07T07:37:00.4606196Z Entering 'third_party/pybind11' 2025-09-07T07:37:00.4653411Z Entering 'third_party/python-peachpy' 2025-09-07T07:37:00.4701181Z Entering 'third_party/sleef' 2025-09-07T07:37:00.4751944Z Entering 'third_party/tensorpipe' 2025-09-07T07:37:00.4801218Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:37:00.4848804Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:37:00.4895624Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:37:00.4942335Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:37:00.4989288Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:37:00.5057373Z ##[endgroup] 2025-09-07T07:37:00.5057737Z ##[group]Persisting credentials for submodules 2025-09-07T07:37:00.5063402Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'url\.https\:\/\/github\.com\/\.insteadOf' && git config --local --unset-all 'url.https://github.com/.insteadOf' || :" 2025-09-07T07:37:00.5421569Z Entering 'android/libs/fbjni' 2025-09-07T07:37:00.5487191Z Entering 'third_party/FP16' 2025-09-07T07:37:00.5550982Z Entering 'third_party/FXdiv' 2025-09-07T07:37:00.5615090Z Entering 'third_party/NNPACK' 2025-09-07T07:37:00.5678714Z Entering 'third_party/NVTX' 2025-09-07T07:37:00.5743945Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:37:00.5807051Z Entering 'third_party/XNNPACK' 2025-09-07T07:37:00.5884759Z Entering 'third_party/aiter' 2025-09-07T07:37:00.5952934Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:37:00.6027057Z Entering 'third_party/benchmark' 2025-09-07T07:37:00.6093579Z Entering 'third_party/composable_kernel' 2025-09-07T07:37:00.6165040Z Entering 'third_party/cpp-httplib' 2025-09-07T07:37:00.6230961Z Entering 'third_party/cpuinfo' 2025-09-07T07:37:00.6296805Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:37:00.6358968Z Entering 'third_party/cutlass' 2025-09-07T07:37:00.6428827Z Entering 'third_party/fbgemm' 2025-09-07T07:37:00.6493816Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:37:00.6557486Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:37:00.6627906Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:37:00.6691674Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:37:00.6764931Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:37:00.6832651Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:37:00.6897833Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:37:00.6967036Z Entering 'third_party/flash-attention' 2025-09-07T07:37:00.7033262Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:37:00.7101623Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:37:00.7176529Z Entering 'third_party/flatbuffers' 2025-09-07T07:37:00.7243534Z Entering 'third_party/fmt' 2025-09-07T07:37:00.7307239Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:37:00.7370190Z Entering 'third_party/gloo' 2025-09-07T07:37:00.7434331Z Entering 'third_party/googletest' 2025-09-07T07:37:00.7498548Z Entering 'third_party/ideep' 2025-09-07T07:37:00.7563266Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:37:00.7636871Z Entering 'third_party/ittapi' 2025-09-07T07:37:00.7703664Z Entering 'third_party/kineto' 2025-09-07T07:37:00.7770763Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:37:00.7834102Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:37:00.7902520Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:37:00.7968005Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:37:00.8029848Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:37:00.8095647Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:37:00.8159619Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:37:00.8222971Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:37:00.8285881Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:37:00.8353007Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:37:00.8420614Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:37:00.8486959Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:37:00.8556034Z Entering 'third_party/kleidiai' 2025-09-07T07:37:00.8625297Z Entering 'third_party/mimalloc' 2025-09-07T07:37:00.8693464Z Entering 'third_party/nlohmann' 2025-09-07T07:37:00.8760256Z Entering 'third_party/onnx' 2025-09-07T07:37:00.8837068Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:37:00.8901871Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:37:00.8966031Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:37:00.9026414Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:37:00.9089692Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:37:00.9151199Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:37:00.9215667Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:37:00.9276524Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:37:00.9342137Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:37:00.9406532Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:37:00.9475471Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:37:00.9543600Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:37:00.9622974Z Entering 'third_party/pocketfft' 2025-09-07T07:37:00.9685664Z Entering 'third_party/protobuf' 2025-09-07T07:37:00.9750544Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:37:00.9811996Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:37:00.9878196Z Entering 'third_party/psimd' 2025-09-07T07:37:00.9940797Z Entering 'third_party/pthreadpool' 2025-09-07T07:37:01.0005112Z Entering 'third_party/pybind11' 2025-09-07T07:37:01.0071402Z Entering 'third_party/python-peachpy' 2025-09-07T07:37:01.0140067Z Entering 'third_party/sleef' 2025-09-07T07:37:01.0206474Z Entering 'third_party/tensorpipe' 2025-09-07T07:37:01.0272927Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:37:01.0338752Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:37:01.0402105Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:37:01.0466362Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:37:01.0530281Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:37:01.0616146Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local 'http.https://github.com/.extraheader' 'AUTHORIZATION: basic ***' && git config --local --show-origin --name-only --get-regexp remote.origin.url" 2025-09-07T07:37:01.0961883Z Entering 'android/libs/fbjni' 2025-09-07T07:37:01.1024752Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-09-07T07:37:01.1045403Z Entering 'third_party/FP16' 2025-09-07T07:37:01.1108832Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-09-07T07:37:01.1128076Z Entering 'third_party/FXdiv' 2025-09-07T07:37:01.1187482Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-09-07T07:37:01.1207658Z Entering 'third_party/NNPACK' 2025-09-07T07:37:01.1269993Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-09-07T07:37:01.1288936Z Entering 'third_party/NVTX' 2025-09-07T07:37:01.1350934Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-09-07T07:37:01.1369894Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:37:01.1428032Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-09-07T07:37:01.1447161Z Entering 'third_party/XNNPACK' 2025-09-07T07:37:01.1504821Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-09-07T07:37:01.1535847Z Entering 'third_party/aiter' 2025-09-07T07:37:01.1594613Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-09-07T07:37:01.1615835Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:37:01.1673017Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-09-07T07:37:01.1701889Z Entering 'third_party/benchmark' 2025-09-07T07:37:01.1763943Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-09-07T07:37:01.1783224Z Entering 'third_party/composable_kernel' 2025-09-07T07:37:01.1846008Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-09-07T07:37:01.1872314Z Entering 'third_party/cpp-httplib' 2025-09-07T07:37:01.1932821Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-09-07T07:37:01.1952909Z Entering 'third_party/cpuinfo' 2025-09-07T07:37:01.2016718Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-09-07T07:37:01.2035833Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:37:01.2096739Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-09-07T07:37:01.2116401Z Entering 'third_party/cutlass' 2025-09-07T07:37:01.2172686Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-09-07T07:37:01.2199583Z Entering 'third_party/fbgemm' 2025-09-07T07:37:01.2257719Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-09-07T07:37:01.2279431Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:37:01.2339432Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-09-07T07:37:01.2356840Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:37:01.2415174Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-09-07T07:37:01.2438927Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:37:01.2499411Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-09-07T07:37:01.2521713Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:37:01.2581105Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-09-07T07:37:01.2608376Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:37:01.2668447Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-09-07T07:37:01.2688072Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:37:01.2747907Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-09-07T07:37:01.2766702Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:37:01.2827598Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-09-07T07:37:01.2851445Z Entering 'third_party/flash-attention' 2025-09-07T07:37:01.2910825Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-09-07T07:37:01.2931411Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:37:01.2989323Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-09-07T07:37:01.3013771Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:37:01.3073409Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-09-07T07:37:01.3100926Z Entering 'third_party/flatbuffers' 2025-09-07T07:37:01.3158372Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-09-07T07:37:01.3180363Z Entering 'third_party/fmt' 2025-09-07T07:37:01.3237918Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-09-07T07:37:01.3258682Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:37:01.3317543Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-09-07T07:37:01.3337645Z Entering 'third_party/gloo' 2025-09-07T07:37:01.3396400Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-09-07T07:37:01.3417892Z Entering 'third_party/googletest' 2025-09-07T07:37:01.3477797Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-09-07T07:37:01.3499146Z Entering 'third_party/ideep' 2025-09-07T07:37:01.3560335Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-09-07T07:37:01.3579873Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:37:01.3639827Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-09-07T07:37:01.3667625Z Entering 'third_party/ittapi' 2025-09-07T07:37:01.3729001Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-09-07T07:37:01.3748837Z Entering 'third_party/kineto' 2025-09-07T07:37:01.3810506Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-09-07T07:37:01.3831008Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:37:01.3890080Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-09-07T07:37:01.3906820Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:37:01.3964122Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-09-07T07:37:01.3984471Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:37:01.4042794Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-09-07T07:37:01.4061845Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:37:01.4122695Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-09-07T07:37:01.4141651Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:37:01.4200437Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-09-07T07:37:01.4219461Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:37:01.4280475Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-09-07T07:37:01.4302195Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:37:01.4361448Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-09-07T07:37:01.4382115Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:37:01.4442961Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-09-07T07:37:01.4463856Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:37:01.4526600Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-09-07T07:37:01.4546121Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:37:01.4607554Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-09-07T07:37:01.4630166Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:37:01.4687291Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-09-07T07:37:01.4706416Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:37:01.4762513Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-09-07T07:37:01.4784083Z Entering 'third_party/kleidiai' 2025-09-07T07:37:01.4846116Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-09-07T07:37:01.4866283Z Entering 'third_party/mimalloc' 2025-09-07T07:37:01.4927095Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-09-07T07:37:01.4946068Z Entering 'third_party/nlohmann' 2025-09-07T07:37:01.5004303Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-09-07T07:37:01.5025311Z Entering 'third_party/onnx' 2025-09-07T07:37:01.5087991Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-09-07T07:37:01.5121372Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:37:01.5179954Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-09-07T07:37:01.5204374Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:37:01.5265632Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-09-07T07:37:01.5288109Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:37:01.5349335Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-09-07T07:37:01.5369562Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:37:01.5430927Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-09-07T07:37:01.5449421Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:37:01.5510931Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-09-07T07:37:01.5529365Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:37:01.5586749Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-09-07T07:37:01.5609250Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:37:01.5666314Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-09-07T07:37:01.5684280Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:37:01.5741514Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-09-07T07:37:01.5759251Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:37:01.5816837Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-09-07T07:37:01.5834549Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:37:01.5892164Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-09-07T07:37:01.5913698Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:37:01.5974394Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-09-07T07:37:01.5997750Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:37:01.6059745Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-09-07T07:37:01.6096576Z Entering 'third_party/pocketfft' 2025-09-07T07:37:01.6159555Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-09-07T07:37:01.6179722Z Entering 'third_party/protobuf' 2025-09-07T07:37:01.6241142Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-09-07T07:37:01.6264611Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:37:01.6321869Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-09-07T07:37:01.6342332Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:37:01.6403652Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-09-07T07:37:01.6425126Z Entering 'third_party/psimd' 2025-09-07T07:37:01.6483752Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-09-07T07:37:01.6502851Z Entering 'third_party/pthreadpool' 2025-09-07T07:37:01.6562249Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-09-07T07:37:01.6580663Z Entering 'third_party/pybind11' 2025-09-07T07:37:01.6638756Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-09-07T07:37:01.6659368Z Entering 'third_party/python-peachpy' 2025-09-07T07:37:01.6719770Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-09-07T07:37:01.6740556Z Entering 'third_party/sleef' 2025-09-07T07:37:01.6800065Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-09-07T07:37:01.6821135Z Entering 'third_party/tensorpipe' 2025-09-07T07:37:01.6881901Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-09-07T07:37:01.6901423Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:37:01.6962723Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-09-07T07:37:01.6984155Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:37:01.7045117Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-09-07T07:37:01.7062404Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:37:01.7121776Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-09-07T07:37:01.7140384Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:37:01.7198265Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-09-07T07:37:01.7219934Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:37:01.7277400Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-09-07T07:37:01.7840097Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'git@github.com:' 2025-09-07T07:37:01.8201405Z Entering 'android/libs/fbjni' 2025-09-07T07:37:01.8250732Z Entering 'third_party/FP16' 2025-09-07T07:37:01.8302046Z Entering 'third_party/FXdiv' 2025-09-07T07:37:01.8349947Z Entering 'third_party/NNPACK' 2025-09-07T07:37:01.8400664Z Entering 'third_party/NVTX' 2025-09-07T07:37:01.8452319Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:37:01.8500793Z Entering 'third_party/XNNPACK' 2025-09-07T07:37:01.8565234Z Entering 'third_party/aiter' 2025-09-07T07:37:01.8617589Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:37:01.8673796Z Entering 'third_party/benchmark' 2025-09-07T07:37:01.8724747Z Entering 'third_party/composable_kernel' 2025-09-07T07:37:01.8783617Z Entering 'third_party/cpp-httplib' 2025-09-07T07:37:01.8833018Z Entering 'third_party/cpuinfo' 2025-09-07T07:37:01.8880893Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:37:01.8928272Z Entering 'third_party/cutlass' 2025-09-07T07:37:01.8983058Z Entering 'third_party/fbgemm' 2025-09-07T07:37:01.9036406Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:37:01.9081848Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:37:01.9134735Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:37:01.9181699Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:37:01.9238642Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:37:01.9289436Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:37:01.9340576Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:37:01.9394568Z Entering 'third_party/flash-attention' 2025-09-07T07:37:01.9447120Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:37:01.9500293Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:37:01.9560401Z Entering 'third_party/flatbuffers' 2025-09-07T07:37:01.9612125Z Entering 'third_party/fmt' 2025-09-07T07:37:01.9658881Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:37:01.9707532Z Entering 'third_party/gloo' 2025-09-07T07:37:01.9755129Z Entering 'third_party/googletest' 2025-09-07T07:37:01.9804608Z Entering 'third_party/ideep' 2025-09-07T07:37:01.9852023Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:37:01.9905578Z Entering 'third_party/ittapi' 2025-09-07T07:37:01.9952579Z Entering 'third_party/kineto' 2025-09-07T07:37:02.0001017Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:37:02.0049470Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:37:02.0097786Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:37:02.0148478Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:37:02.0196082Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:37:02.0247036Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:37:02.0297917Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:37:02.0348475Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:37:02.0398975Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:37:02.0452662Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:37:02.0504951Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:37:02.0551531Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:37:02.0603265Z Entering 'third_party/kleidiai' 2025-09-07T07:37:02.0653918Z Entering 'third_party/mimalloc' 2025-09-07T07:37:02.0701454Z Entering 'third_party/nlohmann' 2025-09-07T07:37:02.0751935Z Entering 'third_party/onnx' 2025-09-07T07:37:02.0812961Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:37:02.0864295Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:37:02.0915969Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:37:02.0963999Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:37:02.1012413Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:37:02.1061084Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:37:02.1111061Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:37:02.1160012Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:37:02.1210483Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:37:02.1260292Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:37:02.1307912Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:37:02.1358663Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:37:02.1422087Z Entering 'third_party/pocketfft' 2025-09-07T07:37:02.1468967Z Entering 'third_party/protobuf' 2025-09-07T07:37:02.1519314Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:37:02.1564867Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:37:02.1615356Z Entering 'third_party/psimd' 2025-09-07T07:37:02.1661986Z Entering 'third_party/pthreadpool' 2025-09-07T07:37:02.1711604Z Entering 'third_party/pybind11' 2025-09-07T07:37:02.1762409Z Entering 'third_party/python-peachpy' 2025-09-07T07:37:02.1812762Z Entering 'third_party/sleef' 2025-09-07T07:37:02.1862540Z Entering 'third_party/tensorpipe' 2025-09-07T07:37:02.1912891Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:37:02.1960820Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:37:02.2011387Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:37:02.2059187Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:37:02.2105472Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:37:02.2179349Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'org-21003710@github.com:' 2025-09-07T07:37:02.2527375Z Entering 'android/libs/fbjni' 2025-09-07T07:37:02.2576880Z Entering 'third_party/FP16' 2025-09-07T07:37:02.2628467Z Entering 'third_party/FXdiv' 2025-09-07T07:37:02.2678981Z Entering 'third_party/NNPACK' 2025-09-07T07:37:02.2731066Z Entering 'third_party/NVTX' 2025-09-07T07:37:02.2783654Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:37:02.2836490Z Entering 'third_party/XNNPACK' 2025-09-07T07:37:02.2898778Z Entering 'third_party/aiter' 2025-09-07T07:37:02.2949897Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:37:02.3009016Z Entering 'third_party/benchmark' 2025-09-07T07:37:02.3056801Z Entering 'third_party/composable_kernel' 2025-09-07T07:37:02.3113499Z Entering 'third_party/cpp-httplib' 2025-09-07T07:37:02.3162140Z Entering 'third_party/cpuinfo' 2025-09-07T07:37:02.3217118Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:37:02.3264762Z Entering 'third_party/cutlass' 2025-09-07T07:37:02.3321133Z Entering 'third_party/fbgemm' 2025-09-07T07:37:02.3371823Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:37:02.3418937Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:37:02.3472989Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:37:02.3522078Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:37:02.3578772Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:37:02.3629381Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:37:02.3677890Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:37:02.3730226Z Entering 'third_party/flash-attention' 2025-09-07T07:37:02.3781480Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:37:02.3836828Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:37:02.3892095Z Entering 'third_party/flatbuffers' 2025-09-07T07:37:02.3944798Z Entering 'third_party/fmt' 2025-09-07T07:37:02.3991439Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:37:02.4039523Z Entering 'third_party/gloo' 2025-09-07T07:37:02.4086646Z Entering 'third_party/googletest' 2025-09-07T07:37:02.4133867Z Entering 'third_party/ideep' 2025-09-07T07:37:02.4181933Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:37:02.4234635Z Entering 'third_party/ittapi' 2025-09-07T07:37:02.4281874Z Entering 'third_party/kineto' 2025-09-07T07:37:02.4335513Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:37:02.4382895Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:37:02.4434607Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:37:02.4485019Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:37:02.4534819Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:37:02.4583485Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:37:02.4640470Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:37:02.4688468Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:37:02.4740231Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:37:02.4787343Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:37:02.4837784Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:37:02.4884510Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:37:02.4934465Z Entering 'third_party/kleidiai' 2025-09-07T07:37:02.4982924Z Entering 'third_party/mimalloc' 2025-09-07T07:37:02.5032868Z Entering 'third_party/nlohmann' 2025-09-07T07:37:02.5085333Z Entering 'third_party/onnx' 2025-09-07T07:37:02.5147045Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:37:02.5200197Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:37:02.5254477Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:37:02.5301071Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:37:02.5352640Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:37:02.5402763Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:37:02.5458088Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:37:02.5505389Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:37:02.5551863Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:37:02.5597503Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:37:02.5649006Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:37:02.5699283Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:37:02.5763497Z Entering 'third_party/pocketfft' 2025-09-07T07:37:02.5811849Z Entering 'third_party/protobuf' 2025-09-07T07:37:02.5863602Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:37:02.5910306Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:37:02.5964178Z Entering 'third_party/psimd' 2025-09-07T07:37:02.6015450Z Entering 'third_party/pthreadpool' 2025-09-07T07:37:02.6064690Z Entering 'third_party/pybind11' 2025-09-07T07:37:02.6115436Z Entering 'third_party/python-peachpy' 2025-09-07T07:37:02.6165155Z Entering 'third_party/sleef' 2025-09-07T07:37:02.6216608Z Entering 'third_party/tensorpipe' 2025-09-07T07:37:02.6267595Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:37:02.6314480Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:37:02.6362150Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:37:02.6412093Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:37:02.6457598Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:37:02.6522788Z ##[endgroup] 2025-09-07T07:37:02.6562577Z [command]/usr/bin/git log -1 --format=%H 2025-09-07T07:37:02.6588594Z 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:37:02.6688800Z ##[group]Run cd "${GITHUB_WORKSPACE}" 2025-09-07T07:37:02.6689049Z cd "${GITHUB_WORKSPACE}" 2025-09-07T07:37:02.6689248Z # Clean stale submodule dirs 2025-09-07T07:37:02.6689451Z if [ -z "${NO_SUDO}" ]; then 2025-09-07T07:37:02.6689701Z  sudo git submodule foreach --recursive git clean -ffdx 2025-09-07T07:37:02.6689943Z else 2025-09-07T07:37:02.6690283Z  git submodule foreach --recursive git clean -ffdx 2025-09-07T07:37:02.6690510Z fi 2025-09-07T07:37:02.6702610Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:02.6702846Z env: 2025-09-07T07:37:02.6702990Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:02.6703373Z NO_SUDO: true 2025-09-07T07:37:02.6703519Z ##[endgroup] 2025-09-07T07:37:02.7092771Z Entering 'android/libs/fbjni' 2025-09-07T07:37:02.7134182Z Entering 'third_party/FP16' 2025-09-07T07:37:02.7171873Z Entering 'third_party/FXdiv' 2025-09-07T07:37:02.7209165Z Entering 'third_party/NNPACK' 2025-09-07T07:37:02.7253432Z Entering 'third_party/NVTX' 2025-09-07T07:37:02.7295786Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T07:37:02.7332562Z Entering 'third_party/XNNPACK' 2025-09-07T07:37:02.7452952Z Entering 'third_party/aiter' 2025-09-07T07:37:02.7499881Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T07:37:02.7619585Z Entering 'third_party/benchmark' 2025-09-07T07:37:02.7659341Z Entering 'third_party/composable_kernel' 2025-09-07T07:37:02.7787516Z Entering 'third_party/cpp-httplib' 2025-09-07T07:37:02.7826591Z Entering 'third_party/cpuinfo' 2025-09-07T07:37:02.7871231Z Entering 'third_party/cudnn_frontend' 2025-09-07T07:37:02.7915548Z Entering 'third_party/cutlass' 2025-09-07T07:37:02.8017576Z Entering 'third_party/fbgemm' 2025-09-07T07:37:02.8085058Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T07:37:02.8121649Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T07:37:02.8236502Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T07:37:02.8274661Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T07:37:02.8377367Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T07:37:02.8416082Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T07:37:02.8449543Z Entering 'third_party/fbgemm/external/json' 2025-09-07T07:37:02.8500617Z Entering 'third_party/flash-attention' 2025-09-07T07:37:02.8547625Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T07:37:02.8655594Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T07:37:02.8752140Z Entering 'third_party/flatbuffers' 2025-09-07T07:37:02.8832988Z Entering 'third_party/fmt' 2025-09-07T07:37:02.8870985Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T07:37:02.8913240Z Entering 'third_party/gloo' 2025-09-07T07:37:02.8950693Z Entering 'third_party/googletest' 2025-09-07T07:37:02.8995742Z Entering 'third_party/ideep' 2025-09-07T07:37:02.9034318Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T07:37:02.9122863Z Entering 'third_party/ittapi' 2025-09-07T07:37:02.9161237Z Entering 'third_party/kineto' 2025-09-07T07:37:02.9201258Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T07:37:02.9240595Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T07:37:02.9291134Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T07:37:02.9328290Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T07:37:02.9366399Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T07:37:02.9402663Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T07:37:02.9444193Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T07:37:02.9481449Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T07:37:02.9524709Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T07:37:02.9571596Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T07:37:02.9613220Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T07:37:02.9651117Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T07:37:02.9692632Z Entering 'third_party/kleidiai' 2025-09-07T07:37:02.9734800Z Entering 'third_party/mimalloc' 2025-09-07T07:37:02.9774662Z Entering 'third_party/nlohmann' 2025-09-07T07:37:02.9828845Z Entering 'third_party/onnx' 2025-09-07T07:37:03.0190238Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T07:37:03.0237755Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T07:37:03.0304709Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T07:37:03.0340284Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T07:37:03.0382782Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T07:37:03.0420188Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T07:37:03.0469910Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T07:37:03.0505318Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T07:37:03.0541034Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T07:37:03.0580089Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T07:37:03.0635199Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T07:37:03.0674501Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T07:37:03.0953602Z Entering 'third_party/pocketfft' 2025-09-07T07:37:03.0992519Z Entering 'third_party/protobuf' 2025-09-07T07:37:03.1079456Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T07:37:03.1118077Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T07:37:03.1162121Z Entering 'third_party/psimd' 2025-09-07T07:37:03.1202230Z Entering 'third_party/pthreadpool' 2025-09-07T07:37:03.1239626Z Entering 'third_party/pybind11' 2025-09-07T07:37:03.1280300Z Entering 'third_party/python-peachpy' 2025-09-07T07:37:03.1321355Z Entering 'third_party/sleef' 2025-09-07T07:37:03.1361262Z Entering 'third_party/tensorpipe' 2025-09-07T07:37:03.1400641Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T07:37:03.1440592Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T07:37:03.1474276Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T07:37:03.1517208Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T07:37:03.1551764Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T07:37:03.1686402Z Prepare all required actions 2025-09-07T07:37:03.1686836Z Getting action download info 2025-09-07T07:37:03.2795578Z ##[group]Run ./.github/actions/setup-linux 2025-09-07T07:37:03.2795798Z env: 2025-09-07T07:37:03.2795955Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:03.2796122Z ##[endgroup] 2025-09-07T07:37:03.2826944Z ##[group]Run set -euo pipefail 2025-09-07T07:37:03.2827177Z set -euo pipefail 2025-09-07T07:37:03.2827363Z function get_ec2_metadata() { 2025-09-07T07:37:03.2827609Z  # Pulled from instance metadata endpoint for EC2 2025-09-07T07:37:03.2828021Z  # see https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/instancedata-data-retrieval.html 2025-09-07T07:37:03.2828369Z  category=$1 2025-09-07T07:37:03.2828602Z  # If it is GCP runner (runner name contains gcp), do not run this 2025-09-07T07:37:03.2828873Z  runner_name_str=i-0d9447236daff7b96 2025-09-07T07:37:03.2829109Z  if [[ -f /.inarc ]]; then 2025-09-07T07:37:03.2829330Z  echo "ARC Runner, no info on ec2 metadata" 2025-09-07T07:37:03.2829571Z  elif [[ $runner_name_str == *"gcp"* ]]; then 2025-09-07T07:37:03.2829855Z  echo "Runner is from Google Cloud Platform, No info on ec2 metadata" 2025-09-07T07:37:03.2830113Z  else 2025-09-07T07:37:03.2830635Z  curl -H "X-aws-ec2-metadata-token: $(curl -s -X PUT "http://169.254.169.254/latest/api/token" -H "X-aws-ec2-metadata-token-ttl-seconds: 30")" -fsSL "http://169.254.169.254/latest/meta-data/${category}" 2025-09-07T07:37:03.2831160Z  fi 2025-09-07T07:37:03.2831301Z } 2025-09-07T07:37:03.2831482Z echo "ami-id: $(get_ec2_metadata ami-id)" 2025-09-07T07:37:03.2831745Z echo "instance-id: $(get_ec2_metadata instance-id)" 2025-09-07T07:37:03.2832039Z echo "instance-type: $(get_ec2_metadata instance-type)" 2025-09-07T07:37:03.2832289Z echo "system info $(uname -a)" 2025-09-07T07:37:03.2840522Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:03.2840761Z env: 2025-09-07T07:37:03.2840908Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:03.2841075Z ##[endgroup] 2025-09-07T07:37:03.2999929Z ami-id: ami-05ffe3c48a9991133 2025-09-07T07:37:03.3115204Z instance-id: i-0d9447236daff7b96 2025-09-07T07:37:03.3220431Z instance-type: m7a.24xlarge 2025-09-07T07:37:03.3233005Z system info Linux ip-10-0-57-198.ec2.internal 6.1.141-155.222.amzn2023.x86_64 #1 SMP PREEMPT_DYNAMIC Tue Jun 17 10:29:47 UTC 2025 x86_64 x86_64 x86_64 GNU/Linux 2025-09-07T07:37:03.3252889Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:37:03.3253469Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:37:03.3262201Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:03.3262459Z env: 2025-09-07T07:37:03.3262611Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:03.3262782Z ##[endgroup] 2025-09-07T07:37:03.3334888Z ##[group]Run if systemctl is-active --quiet docker; then 2025-09-07T07:37:03.3335191Z if systemctl is-active --quiet docker; then 2025-09-07T07:37:03.3335438Z  echo "Docker daemon is running..."; 2025-09-07T07:37:03.3335654Z else 2025-09-07T07:37:03.3335875Z  echo "Starting docker daemon..." && sudo systemctl start docker; 2025-09-07T07:37:03.3336139Z fi 2025-09-07T07:37:03.3343316Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:03.3343728Z env: 2025-09-07T07:37:03.3343871Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:03.3344044Z ##[endgroup] 2025-09-07T07:37:03.3429845Z Docker daemon is running... 2025-09-07T07:37:03.3462426Z ##[group]Run nick-fields/retry@v3.0.0 2025-09-07T07:37:03.3462646Z with: 2025-09-07T07:37:03.3462778Z shell: bash 2025-09-07T07:37:03.3463123Z timeout_minutes: 5 2025-09-07T07:37:03.3463289Z max_attempts: 3 2025-09-07T07:37:03.3463454Z retry_wait_seconds: 30 2025-09-07T07:37:03.3464826Z command: AWS_ACCOUNT_ID=$(aws sts get-caller-identity|grep Account|cut -f4 -d\") aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" # For LF Runners we need to make sure we also login to Meta's ECR docker registry too. META_AWS_ACCOUNT_ID=308535385114 if [ "$AWS_ACCOUNT_ID" != "$META_AWS_ACCOUNT_ID" ] ; then aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$META_AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" fi 2025-09-07T07:37:03.3466148Z polling_interval_seconds: 1 2025-09-07T07:37:03.3466333Z warning_on_retry: true 2025-09-07T07:37:03.3466506Z continue_on_error: false 2025-09-07T07:37:03.3466665Z env: 2025-09-07T07:37:03.3466809Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:03.3466977Z AWS_RETRY_MODE: standard 2025-09-07T07:37:03.3467143Z AWS_MAX_ATTEMPTS: 5 2025-09-07T07:37:03.3467304Z AWS_DEFAULT_REGION: us-east-1 2025-09-07T07:37:03.3467481Z ##[endgroup] 2025-09-07T07:37:04.3632290Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-09-07T07:37:04.3632703Z Configure a credential helper to remove this warning. See 2025-09-07T07:37:04.3633075Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-09-07T07:37:04.3633314Z 2025-09-07T07:37:04.3633398Z Login Succeeded 2025-09-07T07:37:04.5188871Z Command completed after 1 attempt(s). 2025-09-07T07:37:04.5243115Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:37:04.5243460Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:37:04.5243748Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:37:04.5253669Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:04.5253902Z env: 2025-09-07T07:37:04.5254052Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:04.5254225Z ##[endgroup] 2025-09-07T07:37:04.5347201Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-09-07T07:37:04.5347581Z # ignore expansion of "docker ps -q" since it could be empty 2025-09-07T07:37:04.5347853Z # shellcheck disable=SC2046 2025-09-07T07:37:04.5379437Z docker stop $(docker ps -q) || true 2025-09-07T07:37:04.5379757Z # Prune all of the docker images 2025-09-07T07:37:04.5379997Z docker system prune -af 2025-09-07T07:37:04.5387432Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:04.5387662Z env: 2025-09-07T07:37:04.5387801Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:04.5387961Z ##[endgroup] 2025-09-07T07:37:04.5884831Z "docker stop" requires at least 1 argument. 2025-09-07T07:37:04.5885120Z See 'docker stop --help'. 2025-09-07T07:37:04.5885267Z 2025-09-07T07:37:04.5885382Z Usage: docker stop [OPTIONS] CONTAINER [CONTAINER...] 2025-09-07T07:37:04.5885560Z 2025-09-07T07:37:04.5885640Z Stop one or more running containers 2025-09-07T07:37:04.6243258Z Total reclaimed space: 0B 2025-09-07T07:37:04.6278734Z ##[group]Run set +e 2025-09-07T07:37:04.6278944Z set +e 2025-09-07T07:37:04.6279098Z set -x 2025-09-07T07:37:04.6279240Z  2025-09-07T07:37:04.6279403Z PT_DOMAIN=download.pytorch.org 2025-09-07T07:37:04.6279758Z # TODO: Flaky access to download.pytorch.org https://github.com/pytorch/pytorch/issues/100400, 2025-09-07T07:37:04.6280397Z # cleaning this up once the issue is fixed. There are more than one resolved IP here, the last 2025-09-07T07:37:04.6280715Z # one is returned at random 2025-09-07T07:37:04.6280966Z RESOLVED_IP=$(dig -4 +short "${PT_DOMAIN}" | tail -n1) 2025-09-07T07:37:04.6281201Z  2025-09-07T07:37:04.6281497Z if [ -z "${RESOLVED_IP}" ]; then 2025-09-07T07:37:04.6281769Z  echo "Couldn't resolve ${PT_DOMAIN}, retrying with Google DNS..." 2025-09-07T07:37:04.6282084Z  RESOLVED_IP=$(dig -4 +short "${PT_DOMAIN}" @8.8.8.8 | tail -n1) 2025-09-07T07:37:04.6282319Z  2025-09-07T07:37:04.6282473Z  if [ -z "${RESOLVED_IP}" ]; then 2025-09-07T07:37:04.6282702Z  echo "Couldn't resolve ${PT_DOMAIN}, exiting..." 2025-09-07T07:37:04.6282924Z  exit 1 2025-09-07T07:37:04.6283074Z  fi 2025-09-07T07:37:04.6283214Z fi 2025-09-07T07:37:04.6283340Z  2025-09-07T07:37:04.6283507Z if grep -r "${PT_DOMAIN}" /etc/hosts; then 2025-09-07T07:37:04.6283735Z  # Clean up any old records first 2025-09-07T07:37:04.6283955Z  sudo sed -i "/${PT_DOMAIN}/d" /etc/hosts 2025-09-07T07:37:04.6284144Z fi 2025-09-07T07:37:04.6284276Z  2025-09-07T07:37:04.6284473Z echo "${RESOLVED_IP} ${PT_DOMAIN}" | sudo tee -a /etc/hosts 2025-09-07T07:37:04.6284713Z cat /etc/hosts 2025-09-07T07:37:04.6291944Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:04.6292181Z env: 2025-09-07T07:37:04.6292331Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:04.6292503Z ##[endgroup] 2025-09-07T07:37:04.6319974Z + PT_DOMAIN=download.pytorch.org 2025-09-07T07:37:04.6327331Z ++ dig -4 +short download.pytorch.org 2025-09-07T07:37:04.6328102Z ++ tail -n1 2025-09-07T07:37:04.6774999Z + RESOLVED_IP=18.160.10.28 2025-09-07T07:37:04.6776038Z + '[' -z 18.160.10.28 ']' 2025-09-07T07:37:04.6776320Z + grep -r download.pytorch.org /etc/hosts 2025-09-07T07:37:04.6796513Z + echo '18.160.10.28 download.pytorch.org' 2025-09-07T07:37:04.6797080Z + sudo tee -a /etc/hosts 2025-09-07T07:37:05.0761666Z 18.160.10.28 download.pytorch.org 2025-09-07T07:37:05.0778497Z + cat /etc/hosts 2025-09-07T07:37:05.0792016Z 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 2025-09-07T07:37:05.0796731Z ::1 localhost6 localhost6.localdomain6 2025-09-07T07:37:05.0797009Z 18.160.10.28 download.pytorch.org 2025-09-07T07:37:05.0900730Z ##[group]Run pytorch/test-infra/.github/actions/calculate-docker-image@main 2025-09-07T07:37:05.0901039Z with: 2025-09-07T07:37:05.0901570Z docker-image-name: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:05.0902138Z use-custom-docker-registry: true 2025-09-07T07:37:05.0902335Z docker-build-dir: .ci/docker 2025-09-07T07:37:05.0902524Z docker-build-script: ./build.sh 2025-09-07T07:37:05.0902731Z working-directory: . 2025-09-07T07:37:05.0902973Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:05.0903435Z force-push: false 2025-09-07T07:37:05.0903589Z env: 2025-09-07T07:37:05.0903732Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:05.0903901Z ##[endgroup] 2025-09-07T07:37:05.0917025Z ##[group]Run set -ex 2025-09-07T07:37:05.0917225Z set -ex 2025-09-07T07:37:05.0917369Z  2025-09-07T07:37:05.0917642Z # If the docker build directory or the build script doesn't exist, the action will 2025-09-07T07:37:05.0918039Z # gracefully return the docker image name as it is. Pulling docker image in Linux 2025-09-07T07:37:05.0918375Z # job could then download the pre-built image as usual 2025-09-07T07:37:05.0918794Z if [[ -d "${DOCKER_BUILD_DIR}" ]] && [[ -f "${DOCKER_BUILD_DIR}/${DOCKER_BUILD_SCRIPT}" ]] && [[ "${USE_CUSTOM_DOCKER_REGISTRY}" == "true" ]]; then 2025-09-07T07:37:05.0919327Z  echo "skip=false" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:05.0919540Z else 2025-09-07T07:37:05.0919710Z  echo "skip=true" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:05.0919982Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:05.0920227Z  2025-09-07T07:37:05.0920571Z  echo "Not using custom ECR registry. Either it was not requested or there is no Docker build script in the ${REPO_NAME} repo..." 2025-09-07T07:37:05.0920951Z  exit 0 2025-09-07T07:37:05.0921092Z fi 2025-09-07T07:37:05.0921230Z  2025-09-07T07:37:05.0921450Z if [[ "${DOCKER_IMAGE_NAME}" == *"${DOCKER_REGISTRY}/${REPO_NAME}"* ]]; then 2025-09-07T07:37:05.0921813Z  # The docker image name already includes the ECR prefix and tag, so we can just 2025-09-07T07:37:05.0922149Z  # use it as it is, but first let's extract the tag 2025-09-07T07:37:05.0922441Z  DOCKER_TAG=$(echo "${DOCKER_IMAGE_NAME}" | awk -F '[:,]' '{print $2}') 2025-09-07T07:37:05.0922754Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:05.0923046Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:05.0923293Z else 2025-09-07T07:37:05.0923469Z  if [[ "${DOCKER_IMAGE_NAME}" == *:* ]]; then 2025-09-07T07:37:05.0923699Z  CUSTOM_TAG_PREFIX=${DOCKER_IMAGE_NAME#*:} 2025-09-07T07:37:05.0923943Z  DOCKER_IMAGE_NAME=${DOCKER_IMAGE_NAME%%:*} 2025-09-07T07:37:05.0924148Z  fi 2025-09-07T07:37:05.0924427Z  DOCKER_TAG=${CUSTOM_TAG_PREFIX:+${CUSTOM_TAG_PREFIX}-}$(git rev-parse HEAD:"${DOCKER_BUILD_DIR}") 2025-09-07T07:37:05.0924784Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:05.0925160Z  echo "docker-image=${DOCKER_REGISTRY}/${REPO_NAME}/${DOCKER_IMAGE_NAME}:${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:05.0925572Z  echo "custom-tag-prefix=${CUSTOM_TAG_PREFIX}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:05.0925830Z fi 2025-09-07T07:37:05.0935257Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:05.0935488Z env: 2025-09-07T07:37:05.0935632Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:05.0935808Z REPO_NAME: pytorch 2025-09-07T07:37:05.0936491Z DOCKER_IMAGE_NAME: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:05.0937053Z DOCKER_BUILD_DIR: .ci/docker 2025-09-07T07:37:05.0937245Z DOCKER_BUILD_SCRIPT: ./build.sh 2025-09-07T07:37:05.0937489Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:05.0937817Z USE_CUSTOM_DOCKER_REGISTRY: true 2025-09-07T07:37:05.0938000Z CUSTOM_TAG_PREFIX: 2025-09-07T07:37:05.0938156Z ##[endgroup] 2025-09-07T07:37:05.0966760Z + [[ -d .ci/docker ]] 2025-09-07T07:37:05.0966958Z + [[ -f .ci/docker/./build.sh ]] 2025-09-07T07:37:05.0967156Z + [[ true == \t\r\u\e ]] 2025-09-07T07:37:05.0967316Z + echo skip=false 2025-09-07T07:37:05.0968013Z + [[ 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 == *\3\0\8\5\3\5\3\8\5\1\1\4\.\d\k\r\.\e\c\r\.\u\s\-\e\a\s\t\-\1\.\a\m\a\z\o\n\a\w\s\.\c\o\m\/\p\y\t\o\r\c\h* ]] 2025-09-07T07:37:05.0974812Z ++ echo 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:05.0975937Z ++ awk -F '[:,]' '{print $2}' 2025-09-07T07:37:05.1002695Z + DOCKER_TAG=pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:05.1003473Z + echo docker-tag=pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:05.1004246Z + echo docker-image=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:05.1027506Z ##[group]Run set +e 2025-09-07T07:37:05.1027721Z set +e 2025-09-07T07:37:05.1027873Z set -x 2025-09-07T07:37:05.1028017Z  2025-09-07T07:37:05.1028150Z login() { 2025-09-07T07:37:05.1028451Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-09-07T07:37:05.1028777Z } 2025-09-07T07:37:05.1028915Z  2025-09-07T07:37:05.1029046Z retry () { 2025-09-07T07:37:05.1029225Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-09-07T07:37:05.1029433Z } 2025-09-07T07:37:05.1029564Z  2025-09-07T07:37:05.1029707Z retry login "${DOCKER_REGISTRY}" 2025-09-07T07:37:05.1029899Z  2025-09-07T07:37:05.1030042Z START_TIME=$(date +%s) 2025-09-07T07:37:05.1030233Z # Wait up to 120 minutes 2025-09-07T07:37:05.1030479Z while [[ $(( $(date +%s) - 7200 )) -lt $START_TIME ]]; do 2025-09-07T07:37:05.1030770Z  # Check if image already exists, if it does then skip building it 2025-09-07T07:37:05.1031068Z  if docker manifest inspect "${DOCKER_IMAGE}"; then 2025-09-07T07:37:05.1031292Z  exit 0 2025-09-07T07:37:05.1031443Z  fi 2025-09-07T07:37:05.1031577Z  2025-09-07T07:37:05.1031824Z  # NB: This flag is used by Docker build workflow to push the image to ECR, so we can 2025-09-07T07:37:05.1032212Z  # use this to differentiate between the Docker build and regular build jobs. For the 2025-09-07T07:37:05.1032600Z  # latter, it will wait for the Docker images to become available before continuing 2025-09-07T07:37:05.1032914Z  if [ "${DOCKER_PUSH:-false}" == "true" ]; then 2025-09-07T07:37:05.1033159Z  # It's a Docker build job, let's build the image 2025-09-07T07:37:05.1033380Z  break 2025-09-07T07:37:05.1033533Z  else 2025-09-07T07:37:05.1033750Z  # It's a regular build job, wait for the image to become available 2025-09-07T07:37:05.1033995Z  sleep 300 2025-09-07T07:37:05.1034153Z  fi 2025-09-07T07:37:05.1034294Z done 2025-09-07T07:37:05.1034430Z  2025-09-07T07:37:05.1034642Z # NB: This part requires a full checkout. Otherwise, the merge base will 2025-09-07T07:37:05.1035154Z # be empty. The default action would be to continue rebuild the image 2025-09-07T07:37:05.1035474Z if [[ "$BASE_REVISION" = "$(git rev-parse HEAD)" ]]; then 2025-09-07T07:37:05.1035752Z  # if we're on the base branch then use the parent commit 2025-09-07T07:37:05.1035999Z  MERGE_BASE=$(git rev-parse HEAD~) 2025-09-07T07:37:05.1036189Z else 2025-09-07T07:37:05.1036395Z  # otherwise we're on a PR, so use the most recent base commit 2025-09-07T07:37:05.1036685Z  MERGE_BASE=$(git merge-base HEAD "$BASE_REVISION") 2025-09-07T07:37:05.1036903Z fi 2025-09-07T07:37:05.1037034Z  2025-09-07T07:37:05.1037183Z if [[ -z "${MERGE_BASE}" ]]; then 2025-09-07T07:37:05.1037405Z  echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:05.1037613Z  2025-09-07T07:37:05.1037890Z  echo "Finding merge base only works with full checkout, please set fetch-depth to 0, continuing ..." 2025-09-07T07:37:05.1038215Z  exit 0 2025-09-07T07:37:05.1038360Z fi 2025-09-07T07:37:05.1038495Z  2025-09-07T07:37:05.1038685Z if ! git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}"; then 2025-09-07T07:37:05.1039077Z  echo "Directory '${DOCKER_BUILD_DIR}' not found in commit $MERGE_BASE, you should rebase onto a more recent commit" 2025-09-07T07:37:05.1039417Z  exit 1 2025-09-07T07:37:05.1039558Z fi 2025-09-07T07:37:05.1041151Z  2025-09-07T07:37:05.1041375Z PREVIOUS_DOCKER_TAG=$(git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}") 2025-09-07T07:37:05.1041763Z # If no image exists but the hash is the same as the previous hash then we should error out here 2025-09-07T07:37:05.1042115Z if [[ "${PREVIOUS_DOCKER_TAG}" == "${DOCKER_TAG}" ]]; then 2025-09-07T07:37:05.1042524Z  echo "WARNING: Something has gone wrong and the previous image isn't available for the merge-base of your branch" 2025-09-07T07:37:05.1042974Z  echo " Will re-build docker image to store in local cache, TTS may be longer" 2025-09-07T07:37:05.1043241Z fi 2025-09-07T07:37:05.1043383Z  2025-09-07T07:37:05.1043555Z echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-09-07T07:37:05.1050779Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:05.1051015Z env: 2025-09-07T07:37:05.1051161Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:05.1051342Z DOCKER_BUILD_DIR: .ci/docker 2025-09-07T07:37:05.1051566Z BASE_REVISION: 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:37:05.1052158Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:05.1052873Z DOCKER_TAG: pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:05.1053321Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:05.1053557Z DOCKER_PUSH: 2025-09-07T07:37:05.1053710Z ##[endgroup] 2025-09-07T07:37:05.1077321Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:05.1077603Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:05.1080572Z + aws ecr get-login-password --region us-east-1 2025-09-07T07:37:05.1081838Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:05.5066371Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-09-07T07:37:05.5066802Z Configure a credential helper to remove this warning. See 2025-09-07T07:37:05.5067167Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-09-07T07:37:05.5067393Z 2025-09-07T07:37:05.5067472Z Login Succeeded 2025-09-07T07:37:05.5088481Z ++ date +%s 2025-09-07T07:37:05.5100651Z + START_TIME=1757230625 2025-09-07T07:37:05.5105550Z ++ date +%s 2025-09-07T07:37:05.5117855Z + [[ 1757223425 -lt 1757230625 ]] 2025-09-07T07:37:05.5118449Z + docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:05.7486365Z { 2025-09-07T07:37:05.7486615Z "schemaVersion": 2, 2025-09-07T07:37:05.7486923Z "mediaType": "application/vnd.docker.distribution.manifest.v2+json", 2025-09-07T07:37:05.7487212Z "config": { 2025-09-07T07:37:05.7487435Z "mediaType": "application/vnd.docker.container.image.v1+json", 2025-09-07T07:37:05.7487720Z "size": 30269, 2025-09-07T07:37:05.7487989Z "digest": "sha256:662d8c9dfc7db2f5d004293de4f2b7647941dee4c916479ef082d17fcdfd9c47" 2025-09-07T07:37:05.7488282Z }, 2025-09-07T07:37:05.7488411Z "layers": [ 2025-09-07T07:37:05.7488546Z { 2025-09-07T07:37:05.7488754Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7489009Z "size": 30448359, 2025-09-07T07:37:05.7489291Z "digest": "sha256:e6fdc8487bfe6d764301ef3634bc6c043841dc3ab05ca14f81e69c0f92562d46" 2025-09-07T07:37:05.7489576Z }, 2025-09-07T07:37:05.7489696Z { 2025-09-07T07:37:05.7489894Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7490145Z "size": 1554, 2025-09-07T07:37:05.7490402Z "digest": "sha256:18a5ee5b0e2e283bf6d7b9c4c312b0448c75eff1c43446c22c5139a3aeec97fe" 2025-09-07T07:37:05.7490671Z }, 2025-09-07T07:37:05.7490794Z { 2025-09-07T07:37:05.7491030Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7491672Z "size": 313297813, 2025-09-07T07:37:05.7491943Z "digest": "sha256:572424b92528ee46c84fdf3e9e1f5fd75e302621ad75dcf4257ad06778885094" 2025-09-07T07:37:05.7492235Z }, 2025-09-07T07:37:05.7492360Z { 2025-09-07T07:37:05.7492565Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7492814Z "size": 793, 2025-09-07T07:37:05.7493076Z "digest": "sha256:1c35b7d4b67c6769f59f96a643d69c214c5b00291a4968cdd395eedbce82b9c0" 2025-09-07T07:37:05.7493351Z }, 2025-09-07T07:37:05.7493470Z { 2025-09-07T07:37:05.7493669Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7493914Z "size": 106, 2025-09-07T07:37:05.7494164Z "digest": "sha256:68c20f3c23bb0bddb9b69e6ce2e45bcd5b1fcfd9b37dbe3de26b8a5f0e81ff13" 2025-09-07T07:37:05.7494450Z }, 2025-09-07T07:37:05.7494570Z { 2025-09-07T07:37:05.7494762Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7494997Z "size": 704, 2025-09-07T07:37:05.7495237Z "digest": "sha256:7efa39950d3273a15b20bc5f6659373b2b4eb62e36328d96b289834c48d2e408" 2025-09-07T07:37:05.7495506Z }, 2025-09-07T07:37:05.7495626Z { 2025-09-07T07:37:05.7495811Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7496050Z "size": 1214, 2025-09-07T07:37:05.7496303Z "digest": "sha256:a10eb16a7271e996ea9f1d769ba6bd2ec69358f2a79cf26649595a8cea38275f" 2025-09-07T07:37:05.7496580Z }, 2025-09-07T07:37:05.7496699Z { 2025-09-07T07:37:05.7496882Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7497119Z "size": 485, 2025-09-07T07:37:05.7497350Z "digest": "sha256:7d52cf57965449440c17f257fe4c522f9685019961eaa9853d7c820cfe39f5cc" 2025-09-07T07:37:05.7497667Z }, 2025-09-07T07:37:05.7497779Z { 2025-09-07T07:37:05.7497971Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7498219Z "size": 110343705, 2025-09-07T07:37:05.7498479Z "digest": "sha256:cb6a20fcf4e24ec2e1f72ecf361b26e058f3e6194947a9b3a25312223d43516e" 2025-09-07T07:37:05.7498752Z }, 2025-09-07T07:37:05.7498871Z { 2025-09-07T07:37:05.7499060Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7499298Z "size": 4787, 2025-09-07T07:37:05.7499534Z "digest": "sha256:46fb6a8b3e1d4eac9b3a21577824410003ed38f194b4b1486b747e324b32ef6a" 2025-09-07T07:37:05.7499808Z }, 2025-09-07T07:37:05.7500072Z { 2025-09-07T07:37:05.7500263Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7500492Z "size": 1709, 2025-09-07T07:37:05.7500741Z "digest": "sha256:5ad6977cc38e4ea8a6545d6a4fc0e2fdde705a7af96eb496cfe20f264fbc1e74" 2025-09-07T07:37:05.7501020Z }, 2025-09-07T07:37:05.7501138Z { 2025-09-07T07:37:05.7501323Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7501563Z "size": 724, 2025-09-07T07:37:05.7501813Z "digest": "sha256:da63046995a2e510b7146776371a14bff4b31002cc3ef0322e45a3932fba2031" 2025-09-07T07:37:05.7502082Z }, 2025-09-07T07:37:05.7502198Z { 2025-09-07T07:37:05.7502395Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7502642Z "size": 543, 2025-09-07T07:37:05.7502883Z "digest": "sha256:78243fdb9906cb588921ddaa67a3ca915aa9447ca675faac1a9ebc420a561d83" 2025-09-07T07:37:05.7503339Z }, 2025-09-07T07:37:05.7503459Z { 2025-09-07T07:37:05.7503652Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7503895Z "size": 3395447162, 2025-09-07T07:37:05.7504143Z "digest": "sha256:6f70d5d50abaab8988f460b5590d92b6d1d340575ddee981662c24034d7d20af" 2025-09-07T07:37:05.7504415Z }, 2025-09-07T07:37:05.7504536Z { 2025-09-07T07:37:05.7504725Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7504955Z "size": 32, 2025-09-07T07:37:05.7505197Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:05.7505591Z }, 2025-09-07T07:37:05.7505715Z { 2025-09-07T07:37:05.7505905Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7506150Z "size": 380, 2025-09-07T07:37:05.7506402Z "digest": "sha256:69715d3ad3c493436abde51f5a575e79f7d55b46c653f5607f3c7722ad9a05db" 2025-09-07T07:37:05.7506682Z }, 2025-09-07T07:37:05.7506798Z { 2025-09-07T07:37:05.7506995Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7507240Z "size": 235844, 2025-09-07T07:37:05.7507494Z "digest": "sha256:7ace90c063f3f3ce8f04b541afe935088868930e5c074824af2b2c327779a3b5" 2025-09-07T07:37:05.7507765Z }, 2025-09-07T07:37:05.7507882Z { 2025-09-07T07:37:05.7508076Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7508316Z "size": 230, 2025-09-07T07:37:05.7508552Z "digest": "sha256:acbd5447dd1406dab8e46234f6a034a75ad9794f76c24f817b0ecf28b6a69c78" 2025-09-07T07:37:05.7508826Z }, 2025-09-07T07:37:05.7508963Z { 2025-09-07T07:37:05.7509159Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7509540Z "size": 3396092, 2025-09-07T07:37:05.7509798Z "digest": "sha256:744523d9b7f5a3e7abfc646c2d5222e7379024242430b93cb4b8093574e69022" 2025-09-07T07:37:05.7510069Z }, 2025-09-07T07:37:05.7510192Z { 2025-09-07T07:37:05.7510386Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7510635Z "size": 1477, 2025-09-07T07:37:05.7510887Z "digest": "sha256:5bd615a7b945084e11bcb40190f9d6e50367297237146df7b008fa8c668f29c8" 2025-09-07T07:37:05.7511162Z }, 2025-09-07T07:37:05.7511283Z { 2025-09-07T07:37:05.7511469Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7511709Z "size": 482, 2025-09-07T07:37:05.7511964Z "digest": "sha256:f4986a00e3aecf1d56beaada7aba8c49fbb3683db3c99790ab0aa4caaa34f76f" 2025-09-07T07:37:05.7512249Z }, 2025-09-07T07:37:05.7512362Z { 2025-09-07T07:37:05.7512550Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7512787Z "size": 196, 2025-09-07T07:37:05.7513023Z "digest": "sha256:21902f6e4f8cb76c82e755b8fc9f72e1912bf925ab345ab5b4cc2210f4887a64" 2025-09-07T07:37:05.7513286Z }, 2025-09-07T07:37:05.7513402Z { 2025-09-07T07:37:05.7513591Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7513827Z "size": 608, 2025-09-07T07:37:05.7514176Z "digest": "sha256:d80602abf3ccf0c0b527848a403dfde36e1cf1db1416852385feda5c44bf4363" 2025-09-07T07:37:05.7514454Z }, 2025-09-07T07:37:05.7514573Z { 2025-09-07T07:37:05.7514761Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7514992Z "size": 226, 2025-09-07T07:37:05.7515232Z "digest": "sha256:3c51bf0bc362d34a17911f73c5146cbd668c4d1cf1b944cbf40a604d71cd623a" 2025-09-07T07:37:05.7515502Z }, 2025-09-07T07:37:05.7515617Z { 2025-09-07T07:37:05.7515801Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7516039Z "size": 828, 2025-09-07T07:37:05.7516271Z "digest": "sha256:119ab3bceafa6f2cab4b1f71161195139792990263ee8de82230c6284f0ae20a" 2025-09-07T07:37:05.7516540Z }, 2025-09-07T07:37:05.7516652Z { 2025-09-07T07:37:05.7516837Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7517072Z "size": 32, 2025-09-07T07:37:05.7517315Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:05.7517581Z }, 2025-09-07T07:37:05.7517702Z { 2025-09-07T07:37:05.7517890Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7518134Z "size": 104, 2025-09-07T07:37:05.7518370Z "digest": "sha256:af8eadc9eaabdaf6c5e01031d63061605327153e07568ddd159966ecea75cd07" 2025-09-07T07:37:05.7518645Z }, 2025-09-07T07:37:05.7518762Z { 2025-09-07T07:37:05.7518948Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7519259Z "size": 1495, 2025-09-07T07:37:05.7519508Z "digest": "sha256:e7769b0d7a8262f3cc32a9d96080de5318dac3d2617e10508a167e689016e40c" 2025-09-07T07:37:05.7519783Z }, 2025-09-07T07:37:05.7532207Z { 2025-09-07T07:37:05.7532450Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7532729Z "size": 453908015, 2025-09-07T07:37:05.7533014Z "digest": "sha256:ba263639b0f4634277ef3b8903e3457ac27ce012f1bbeeeeb773191c2c3b222b" 2025-09-07T07:37:05.7533310Z }, 2025-09-07T07:37:05.7533435Z { 2025-09-07T07:37:05.7533653Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7533906Z "size": 164, 2025-09-07T07:37:05.7534163Z "digest": "sha256:a5ab7a280382a797dd5ba6a6716f667a231540ad1e0e7c8ba48bb24d5ab80ef0" 2025-09-07T07:37:05.7534445Z }, 2025-09-07T07:37:05.7534565Z { 2025-09-07T07:37:05.7534766Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7535008Z "size": 346, 2025-09-07T07:37:05.7535259Z "digest": "sha256:80b2232d952f55c3662cffd657ba30fe825f08dfcc5bbea13e2bc6de4482b7e4" 2025-09-07T07:37:05.7535532Z }, 2025-09-07T07:37:05.7535653Z { 2025-09-07T07:37:05.7535836Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7536079Z "size": 32, 2025-09-07T07:37:05.7536320Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:05.7536602Z }, 2025-09-07T07:37:05.7536725Z { 2025-09-07T07:37:05.7536911Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7537148Z "size": 106, 2025-09-07T07:37:05.7537391Z "digest": "sha256:cc93cd65e90f0a9c50194579c93e96897f4e582b9777a1c4d7df7b913ddcdded" 2025-09-07T07:37:05.7537737Z }, 2025-09-07T07:37:05.7537852Z { 2025-09-07T07:37:05.7538042Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7538286Z "size": 425, 2025-09-07T07:37:05.7538529Z "digest": "sha256:0eed4c15712bc470dac7df87e33b3570a1510344019dd9cc0e95b8beb1f98372" 2025-09-07T07:37:05.7538795Z }, 2025-09-07T07:37:05.7538919Z { 2025-09-07T07:37:05.7539106Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7539353Z "size": 19309387, 2025-09-07T07:37:05.7539594Z "digest": "sha256:092516f71fe325518f9737f105bcd65c40cd35c3019098889757e2c84c03c8a8" 2025-09-07T07:37:05.7539856Z }, 2025-09-07T07:37:05.7540087Z { 2025-09-07T07:37:05.7540279Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7540512Z "size": 108, 2025-09-07T07:37:05.7540751Z "digest": "sha256:8c0825014a6270f765ff514da8583d55874f3278bef76e5617e29115f91ee654" 2025-09-07T07:37:05.7541021Z }, 2025-09-07T07:37:05.7541141Z { 2025-09-07T07:37:05.7541324Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7541559Z "size": 636, 2025-09-07T07:37:05.7541806Z "digest": "sha256:8e0d2f63da0a8ff07657d7e06cdbc1ad9d5db95614d640a9f7a9aa8c30c9986d" 2025-09-07T07:37:05.7542085Z }, 2025-09-07T07:37:05.7542195Z { 2025-09-07T07:37:05.7542380Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7542618Z "size": 724, 2025-09-07T07:37:05.7542850Z "digest": "sha256:da63046995a2e510b7146776371a14bff4b31002cc3ef0322e45a3932fba2031" 2025-09-07T07:37:05.7543108Z }, 2025-09-07T07:37:05.7543228Z { 2025-09-07T07:37:05.7543419Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7543658Z "size": 148, 2025-09-07T07:37:05.7543886Z "digest": "sha256:73aae7958ba1a16c5f5625d39b06208e1def8c7816bb75028bf0845f553a5068" 2025-09-07T07:37:05.7544154Z }, 2025-09-07T07:37:05.7544273Z { 2025-09-07T07:37:05.7544457Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7544685Z "size": 136, 2025-09-07T07:37:05.7544985Z "digest": "sha256:ac6077ec9fa50fc0822d387d2ee35e1b6f1f56612402fe7195378180b25087bc" 2025-09-07T07:37:05.7545386Z }, 2025-09-07T07:37:05.7545508Z { 2025-09-07T07:37:05.7545690Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7545973Z + exit 0 2025-09-07T07:37:05.7546099Z "size": 140, 2025-09-07T07:37:05.7546341Z "digest": "sha256:bf4ee4e45e92ef179f7fc64e2c7c6755905a969c37cf82c39aafbadd9290ff04" 2025-09-07T07:37:05.7546631Z }, 2025-09-07T07:37:05.7546749Z { 2025-09-07T07:37:05.7546940Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7547183Z "size": 18617175577, 2025-09-07T07:37:05.7547443Z "digest": "sha256:c1b766f9b961bcc863d6f89d623815fd7dfe9797ddcfd5d15ef06ffe7d177359" 2025-09-07T07:37:05.7547717Z }, 2025-09-07T07:37:05.7547833Z { 2025-09-07T07:37:05.7548023Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7548263Z "size": 223, 2025-09-07T07:37:05.7548516Z "digest": "sha256:6e726ef07b5d5cfe2fb9f06d43fc931fc64c381fd37eaf0c169e0dd84796f152" 2025-09-07T07:37:05.7548799Z }, 2025-09-07T07:37:05.7548918Z { 2025-09-07T07:37:05.7549109Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7549350Z "size": 274477524, 2025-09-07T07:37:05.7549593Z "digest": "sha256:364070434a64fa913f3907ada910a4051707e693e0e6124f57bc97aa57791da1" 2025-09-07T07:37:05.7549863Z }, 2025-09-07T07:37:05.7549981Z { 2025-09-07T07:37:05.7550173Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7550410Z "size": 6451569004, 2025-09-07T07:37:05.7550665Z "digest": "sha256:71f708151a84685fc366b85e914dac9f5279313eff07358d79ecaaeecb0f1c42" 2025-09-07T07:37:05.7550929Z }, 2025-09-07T07:37:05.7551049Z { 2025-09-07T07:37:05.7551231Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7551472Z "size": 129, 2025-09-07T07:37:05.7551715Z "digest": "sha256:622d8cfb39ea4dda608d2819c6a9de45df81b6f8319ee8ab4a24c36d81b9a132" 2025-09-07T07:37:05.7551995Z }, 2025-09-07T07:37:05.7552104Z { 2025-09-07T07:37:05.7552291Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7552528Z "size": 778, 2025-09-07T07:37:05.7552762Z "digest": "sha256:284119a92cb13dacff06926444aab4f99756039acb48abba7b75d35c367ed3f1" 2025-09-07T07:37:05.7553025Z }, 2025-09-07T07:37:05.7553147Z { 2025-09-07T07:37:05.7553334Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7553658Z "size": 724, 2025-09-07T07:37:05.7553886Z "digest": "sha256:da63046995a2e510b7146776371a14bff4b31002cc3ef0322e45a3932fba2031" 2025-09-07T07:37:05.7554155Z }, 2025-09-07T07:37:05.7554273Z { 2025-09-07T07:37:05.7554465Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7554697Z "size": 140, 2025-09-07T07:37:05.7554928Z "digest": "sha256:96695940d842555623cfe4fb7b52e949423e8c8f383e55d02363e7e5c5804afa" 2025-09-07T07:37:05.7555193Z }, 2025-09-07T07:37:05.7555317Z { 2025-09-07T07:37:05.7555496Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7555735Z "size": 32, 2025-09-07T07:37:05.7555981Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:05.7556266Z }, 2025-09-07T07:37:05.7556379Z { 2025-09-07T07:37:05.7556570Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7556812Z "size": 160, 2025-09-07T07:37:05.7557056Z "digest": "sha256:7ddca6c4c050460204097ba875dc0fa03eca6265122a18c0b8dc5504152aea53" 2025-09-07T07:37:05.7557323Z }, 2025-09-07T07:37:05.7557440Z { 2025-09-07T07:37:05.7557626Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7557863Z "size": 1012, 2025-09-07T07:37:05.7558122Z "digest": "sha256:a95e1f2f1aadef03514a7cdbdac1fe83d4eebedbb80df9be868a223f27e1c263" 2025-09-07T07:37:05.7558422Z }, 2025-09-07T07:37:05.7558535Z { 2025-09-07T07:37:05.7560266Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7560503Z "size": 724, 2025-09-07T07:37:05.7560735Z "digest": "sha256:da63046995a2e510b7146776371a14bff4b31002cc3ef0322e45a3932fba2031" 2025-09-07T07:37:05.7561000Z }, 2025-09-07T07:37:05.7561118Z { 2025-09-07T07:37:05.7561300Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7561536Z "size": 135, 2025-09-07T07:37:05.7561773Z "digest": "sha256:8085756b0cc0f9588f23a73c27840a5dff48cc18c3a2f0311e4d1ef291855679" 2025-09-07T07:37:05.7562043Z }, 2025-09-07T07:37:05.7562154Z { 2025-09-07T07:37:05.7562340Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7562577Z "size": 32, 2025-09-07T07:37:05.7562820Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:05.7563087Z }, 2025-09-07T07:37:05.7563204Z { 2025-09-07T07:37:05.7563389Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7563627Z "size": 158, 2025-09-07T07:37:05.7563856Z "digest": "sha256:7e9ff0c6f103b18756f01c60b4d57a951660f17bffb1810b330e3ff703caf216" 2025-09-07T07:37:05.7564130Z }, 2025-09-07T07:37:05.7564247Z { 2025-09-07T07:37:05.7564434Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7564668Z "size": 1369, 2025-09-07T07:37:05.7564911Z "digest": "sha256:a625cbbc05b983aeb4c28702a4a5b65c68191ab1b8d17978f7d98cc17ddf3c52" 2025-09-07T07:37:05.7565187Z }, 2025-09-07T07:37:05.7565307Z { 2025-09-07T07:37:05.7565487Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7565724Z "size": 32, 2025-09-07T07:37:05.7565958Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:05.7566230Z }, 2025-09-07T07:37:05.7566345Z { 2025-09-07T07:37:05.7566531Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7566769Z "size": 136, 2025-09-07T07:37:05.7567000Z "digest": "sha256:4e28486424310870c8d6815524440f17c6e0afe7572eaa173a811b98b4920bed" 2025-09-07T07:37:05.7567265Z }, 2025-09-07T07:37:05.7567384Z { 2025-09-07T07:37:05.7567578Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7567817Z "size": 380, 2025-09-07T07:37:05.7568050Z "digest": "sha256:5e944f1ed1bef9442f5b1b86225d3958ea8f2f7f4c6aa7b92dc5d0c810c260bc" 2025-09-07T07:37:05.7568324Z }, 2025-09-07T07:37:05.7568523Z { 2025-09-07T07:37:05.7568717Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7568947Z "size": 32, 2025-09-07T07:37:05.7569194Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:05.7569470Z }, 2025-09-07T07:37:05.7569589Z { 2025-09-07T07:37:05.7569770Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7570004Z "size": 104, 2025-09-07T07:37:05.7570244Z "digest": "sha256:41619248f604c60e038a02bfd462af96ee2996b77be5f59f05e9ac5fe4790e5a" 2025-09-07T07:37:05.7570519Z }, 2025-09-07T07:37:05.7570635Z { 2025-09-07T07:37:05.7570822Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7571051Z "size": 407, 2025-09-07T07:37:05.7571293Z "digest": "sha256:be86f8c4f654b9ae64a20eb7f960e6ce4baa5b46e0a1f5e1312b11492a40bcd4" 2025-09-07T07:37:05.7571571Z }, 2025-09-07T07:37:05.7571686Z { 2025-09-07T07:37:05.7571867Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7572095Z "size": 32, 2025-09-07T07:37:05.7572320Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:05.7572592Z }, 2025-09-07T07:37:05.7572712Z { 2025-09-07T07:37:05.7572897Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7573130Z "size": 109, 2025-09-07T07:37:05.7573364Z "digest": "sha256:ef1340e22a4bc8cf42e1d40961cb32d183cd3da8f0b785b5425c32ee067690c1" 2025-09-07T07:37:05.7573722Z }, 2025-09-07T07:37:05.7573842Z { 2025-09-07T07:37:05.7574042Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7574288Z "size": 1897, 2025-09-07T07:37:05.7574546Z "digest": "sha256:da8d8b696333cbf6b9f339ab859639c905d6752d7e65fea14c23c3c2dcba553e" 2025-09-07T07:37:05.7574819Z }, 2025-09-07T07:37:05.7574939Z { 2025-09-07T07:37:05.7575134Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7575376Z "size": 243443118, 2025-09-07T07:37:05.7575629Z "digest": "sha256:386b0c49c4982a821fb6f427fbc7d9c7d2012e97c96a514a9c7a09304e76b935" 2025-09-07T07:37:05.7575901Z }, 2025-09-07T07:37:05.7576025Z { 2025-09-07T07:37:05.7576218Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7576454Z "size": 106, 2025-09-07T07:37:05.7576714Z "digest": "sha256:2b1d0ea7efe0bf86e86df804d2cddbf83b113fdecd03f3ddfca728da30546f34" 2025-09-07T07:37:05.7577004Z }, 2025-09-07T07:37:05.7577133Z { 2025-09-07T07:37:05.7577314Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7577644Z "size": 163, 2025-09-07T07:37:05.7577890Z "digest": "sha256:04c04be7408f20625b1bd8454e5a08c91fcf04d4f79ab3ec1b75ae6b1824174d" 2025-09-07T07:37:05.7578199Z }, 2025-09-07T07:37:05.7578308Z { 2025-09-07T07:37:05.7578510Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7578793Z "size": 7943, 2025-09-07T07:37:05.7579035Z "digest": "sha256:f8690caa3ac5e845f2dcc25ad12815b5c7452285c3838a87c780bd03ecf072a3" 2025-09-07T07:37:05.7579295Z }, 2025-09-07T07:37:05.7579403Z { 2025-09-07T07:37:05.7579581Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7579807Z "size": 8074, 2025-09-07T07:37:05.7580033Z "digest": "sha256:2908d6baaa6b21331dee5f210472cae0874d22b98b0a35420cad4fd753ed215f" 2025-09-07T07:37:05.7580294Z }, 2025-09-07T07:37:05.7580408Z { 2025-09-07T07:37:05.7580586Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7580817Z "size": 303, 2025-09-07T07:37:05.7581037Z "digest": "sha256:37e2336101eba2c73995d34431e4fae8782d9e9700c42621777922490b2158ed" 2025-09-07T07:37:05.7581289Z }, 2025-09-07T07:37:05.7581398Z { 2025-09-07T07:37:05.7581573Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7581801Z "size": 32, 2025-09-07T07:37:05.7582139Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:05.7582415Z }, 2025-09-07T07:37:05.7582541Z { 2025-09-07T07:37:05.7582754Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7583006Z "size": 108, 2025-09-07T07:37:05.7583232Z "digest": "sha256:f1ac881fde33994861be4324231269058643168b9aee60c699552d0d92d965da" 2025-09-07T07:37:05.7583483Z }, 2025-09-07T07:37:05.7583591Z { 2025-09-07T07:37:05.7583772Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7584002Z "size": 54145699, 2025-09-07T07:37:05.7584239Z "digest": "sha256:43b14c67347e2813c5f63e928c14db60dbb35c330ccc865510cf79739d8b78a1" 2025-09-07T07:37:05.7584496Z }, 2025-09-07T07:37:05.7584604Z { 2025-09-07T07:37:05.7584781Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-09-07T07:37:05.7585005Z "size": 32, 2025-09-07T07:37:05.7585237Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-09-07T07:37:05.7585497Z } 2025-09-07T07:37:05.7585603Z ] 2025-09-07T07:37:05.7585712Z } 2025-09-07T07:37:05.7606630Z ##[group]Run set -eux 2025-09-07T07:37:05.7606822Z set -eux 2025-09-07T07:37:05.7607080Z # It's ok if this steps fails, it would then be an anonymous user like what we used to have 2025-09-07T07:37:05.7607765Z aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token | jq --raw-output '.SecretString' | jq -r .docker_hub_readonly_token | docker login --username pytorchbot --password-stdin || true 2025-09-07T07:37:05.7616923Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:05.7617153Z env: 2025-09-07T07:37:05.7617298Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:05.7617468Z ##[endgroup] 2025-09-07T07:37:05.7646703Z + aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token 2025-09-07T07:37:05.7648376Z + jq --raw-output .SecretString 2025-09-07T07:37:05.7649047Z + jq -r .docker_hub_readonly_token 2025-09-07T07:37:05.7650461Z + docker login --username pytorchbot --password-stdin 2025-09-07T07:37:06.2192682Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-09-07T07:37:06.2193109Z Configure a credential helper to remove this warning. See 2025-09-07T07:37:06.2193488Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-09-07T07:37:06.2193735Z 2025-09-07T07:37:06.2193862Z Login Succeeded 2025-09-07T07:37:06.2277839Z ##[group]Run tag=${ECR_DOCKER_IMAGE##*:} 2025-09-07T07:37:06.2278103Z tag=${ECR_DOCKER_IMAGE##*:} 2025-09-07T07:37:06.2278362Z echo "docker pull ghcr.io/pytorch/ci-image:${tag/:/-}" 2025-09-07T07:37:06.2286620Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:06.2286856Z env: 2025-09-07T07:37:06.2287011Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:06.2287568Z ECR_DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:06.2288104Z ##[endgroup] 2025-09-07T07:37:06.2316580Z docker pull ghcr.io/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:06.2352399Z ##[group]Run pytorch/test-infra/.github/actions/pull-docker-image@main 2025-09-07T07:37:06.2352694Z with: 2025-09-07T07:37:06.2353256Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:06.2353926Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:06.2354181Z env: 2025-09-07T07:37:06.2354338Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:06.2354518Z ##[endgroup] 2025-09-07T07:37:06.2365149Z ##[group]Run set -x 2025-09-07T07:37:06.2365342Z set -x 2025-09-07T07:37:06.2365485Z set +e 2025-09-07T07:37:06.2365625Z  2025-09-07T07:37:06.2365754Z login() { 2025-09-07T07:37:06.2366056Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-09-07T07:37:06.2366364Z } 2025-09-07T07:37:06.2366501Z  2025-09-07T07:37:06.2366873Z retry () { 2025-09-07T07:37:06.2367054Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-09-07T07:37:06.2367248Z } 2025-09-07T07:37:06.2367383Z  2025-09-07T07:37:06.2367543Z retry login "${DOCKER_REGISTRY}" 2025-09-07T07:37:06.2367734Z  2025-09-07T07:37:06.2368029Z IMAGE_SIZE=$(docker manifest inspect "${DOCKER_IMAGE}" | jq '[.layers[].size, .config.size] | add / 1024 / 1024') 2025-09-07T07:37:06.2368423Z echo "Compressed size of image in MB: ${IMAGE_SIZE}" 2025-09-07T07:37:06.2368653Z  2025-09-07T07:37:06.2368789Z set -e 2025-09-07T07:37:06.2369000Z # ignore output since only exit code is used for conditional 2025-09-07T07:37:06.2369302Z # only pull docker image if it's not available locally 2025-09-07T07:37:06.2369629Z if ! docker inspect --type=image "${DOCKER_IMAGE}" >/dev/null 2>/dev/null; then 2025-09-07T07:37:06.2369936Z  retry docker pull "${DOCKER_IMAGE}" 2025-09-07T07:37:06.2370133Z fi 2025-09-07T07:37:06.2377149Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:37:06.2377386Z env: 2025-09-07T07:37:06.2377743Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:37:06.2378282Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:06.2378876Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:06.2379111Z ##[endgroup] 2025-09-07T07:37:06.2405048Z + set +e 2025-09-07T07:37:06.2405273Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:06.2405559Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:06.2409291Z + aws ecr get-login-password --region us-east-1 2025-09-07T07:37:06.2410589Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-09-07T07:37:06.6548096Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-09-07T07:37:06.6548489Z Configure a credential helper to remove this warning. See 2025-09-07T07:37:06.6548846Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-09-07T07:37:06.6549097Z 2025-09-07T07:37:06.6550276Z Login Succeeded 2025-09-07T07:37:06.6576512Z ++ docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:06.6577159Z ++ jq '[.layers[].size, .config.size] | add / 1024 / 1024' 2025-09-07T07:37:06.8644129Z + IMAGE_SIZE=28579.020259857178 2025-09-07T07:37:06.8644437Z + echo 'Compressed size of image in MB: 28579.020259857178' 2025-09-07T07:37:06.8644675Z + set -e 2025-09-07T07:37:06.8645542Z + docker inspect --type=image 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:06.8646163Z Compressed size of image in MB: 28579.020259857178 2025-09-07T07:37:06.8804330Z + retry docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:06.8805240Z + docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:37:07.0876065Z pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77: Pulling from pytorch/ci-image 2025-09-07T07:37:07.0876619Z e6fdc8487bfe: Pulling fs layer 2025-09-07T07:37:07.0876830Z 18a5ee5b0e2e: Pulling fs layer 2025-09-07T07:37:07.0877021Z 572424b92528: Pulling fs layer 2025-09-07T07:37:07.0877196Z 1c35b7d4b67c: Pulling fs layer 2025-09-07T07:37:07.0877360Z 68c20f3c23bb: Pulling fs layer 2025-09-07T07:37:07.0877530Z 7efa39950d32: Pulling fs layer 2025-09-07T07:37:07.0877731Z a10eb16a7271: Pulling fs layer 2025-09-07T07:37:07.0877903Z 7d52cf579654: Pulling fs layer 2025-09-07T07:37:07.0878066Z cb6a20fcf4e2: Pulling fs layer 2025-09-07T07:37:07.0878237Z 46fb6a8b3e1d: Pulling fs layer 2025-09-07T07:37:07.0878421Z 5ad6977cc38e: Pulling fs layer 2025-09-07T07:37:07.0878592Z da63046995a2: Pulling fs layer 2025-09-07T07:37:07.0878752Z 78243fdb9906: Pulling fs layer 2025-09-07T07:37:07.0878918Z 6f70d5d50aba: Pulling fs layer 2025-09-07T07:37:07.0879091Z 4f4fb700ef54: Pulling fs layer 2025-09-07T07:37:07.0879262Z 69715d3ad3c4: Pulling fs layer 2025-09-07T07:37:07.0879422Z 7ace90c063f3: Pulling fs layer 2025-09-07T07:37:07.0879586Z acbd5447dd14: Pulling fs layer 2025-09-07T07:37:07.0879751Z 744523d9b7f5: Pulling fs layer 2025-09-07T07:37:07.0879914Z 5bd615a7b945: Pulling fs layer 2025-09-07T07:37:07.0880077Z f4986a00e3ae: Pulling fs layer 2025-09-07T07:37:07.0880255Z 21902f6e4f8c: Pulling fs layer 2025-09-07T07:37:07.0880422Z d80602abf3cc: Pulling fs layer 2025-09-07T07:37:07.0880591Z 3c51bf0bc362: Pulling fs layer 2025-09-07T07:37:07.0880778Z 119ab3bceafa: Pulling fs layer 2025-09-07T07:37:07.0880949Z af8eadc9eaab: Pulling fs layer 2025-09-07T07:37:07.0881116Z e7769b0d7a82: Pulling fs layer 2025-09-07T07:37:07.0881530Z ba263639b0f4: Pulling fs layer 2025-09-07T07:37:07.0881686Z a5ab7a280382: Pulling fs layer 2025-09-07T07:37:07.0881853Z 80b2232d952f: Pulling fs layer 2025-09-07T07:37:07.0882019Z cc93cd65e90f: Pulling fs layer 2025-09-07T07:37:07.0882185Z 0eed4c15712b: Pulling fs layer 2025-09-07T07:37:07.0882363Z 092516f71fe3: Pulling fs layer 2025-09-07T07:37:07.0882533Z 8c0825014a62: Pulling fs layer 2025-09-07T07:37:07.0882706Z 8e0d2f63da0a: Pulling fs layer 2025-09-07T07:37:07.0882889Z 73aae7958ba1: Pulling fs layer 2025-09-07T07:37:07.0883064Z ac6077ec9fa5: Pulling fs layer 2025-09-07T07:37:07.0883238Z bf4ee4e45e92: Pulling fs layer 2025-09-07T07:37:07.0883414Z c1b766f9b961: Pulling fs layer 2025-09-07T07:37:07.0883581Z 6e726ef07b5d: Pulling fs layer 2025-09-07T07:37:07.0883747Z 364070434a64: Pulling fs layer 2025-09-07T07:37:07.0883908Z 71f708151a84: Pulling fs layer 2025-09-07T07:37:07.0884102Z 622d8cfb39ea: Pulling fs layer 2025-09-07T07:37:07.0884274Z 284119a92cb1: Pulling fs layer 2025-09-07T07:37:07.0884433Z 96695940d842: Pulling fs layer 2025-09-07T07:37:07.0884608Z 7ddca6c4c050: Pulling fs layer 2025-09-07T07:37:07.0884776Z a95e1f2f1aad: Pulling fs layer 2025-09-07T07:37:07.0884941Z 8085756b0cc0: Pulling fs layer 2025-09-07T07:37:07.0885101Z 7e9ff0c6f103: Pulling fs layer 2025-09-07T07:37:07.0885271Z a625cbbc05b9: Pulling fs layer 2025-09-07T07:37:07.0885438Z 4e2848642431: Pulling fs layer 2025-09-07T07:37:07.0885651Z 5e944f1ed1be: Pulling fs layer 2025-09-07T07:37:07.0885818Z 41619248f604: Pulling fs layer 2025-09-07T07:37:07.0885974Z be86f8c4f654: Pulling fs layer 2025-09-07T07:37:07.0886141Z ef1340e22a4b: Pulling fs layer 2025-09-07T07:37:07.0886305Z da8d8b696333: Pulling fs layer 2025-09-07T07:37:07.0886484Z 386b0c49c498: Pulling fs layer 2025-09-07T07:37:07.0888075Z 2b1d0ea7efe0: Pulling fs layer 2025-09-07T07:37:07.0888257Z 04c04be7408f: Pulling fs layer 2025-09-07T07:37:07.0888422Z f8690caa3ac5: Pulling fs layer 2025-09-07T07:37:07.0888582Z 2908d6baaa6b: Pulling fs layer 2025-09-07T07:37:07.0888769Z 37e2336101eb: Pulling fs layer 2025-09-07T07:37:07.0888941Z f1ac881fde33: Pulling fs layer 2025-09-07T07:37:07.0889111Z 43b14c67347e: Pulling fs layer 2025-09-07T07:37:07.0889268Z a95e1f2f1aad: Waiting 2025-09-07T07:37:07.0889429Z 8e0d2f63da0a: Waiting 2025-09-07T07:37:07.0889579Z a10eb16a7271: Waiting 2025-09-07T07:37:07.0889718Z 7d52cf579654: Waiting 2025-09-07T07:37:07.0889862Z acbd5447dd14: Waiting 2025-09-07T07:37:07.0890003Z 46fb6a8b3e1d: Waiting 2025-09-07T07:37:07.0890152Z 5ad6977cc38e: Waiting 2025-09-07T07:37:07.0900324Z 6f70d5d50aba: Waiting 2025-09-07T07:37:07.0900547Z 5bd615a7b945: Waiting 2025-09-07T07:37:07.0900707Z f4986a00e3ae: Waiting 2025-09-07T07:37:07.0900865Z 73aae7958ba1: Waiting 2025-09-07T07:37:07.0901020Z 21902f6e4f8c: Waiting 2025-09-07T07:37:07.0901177Z da63046995a2: Waiting 2025-09-07T07:37:07.0901320Z 0eed4c15712b: Waiting 2025-09-07T07:37:07.0901468Z 69715d3ad3c4: Waiting 2025-09-07T07:37:07.0901611Z 78243fdb9906: Waiting 2025-09-07T07:37:07.0901762Z 1c35b7d4b67c: Waiting 2025-09-07T07:37:07.0901900Z cb6a20fcf4e2: Waiting 2025-09-07T07:37:07.0902043Z 8c0825014a62: Waiting 2025-09-07T07:37:07.0902185Z ba263639b0f4: Waiting 2025-09-07T07:37:07.0902322Z bf4ee4e45e92: Waiting 2025-09-07T07:37:07.0902466Z a5ab7a280382: Waiting 2025-09-07T07:37:07.0902609Z ac6077ec9fa5: Waiting 2025-09-07T07:37:07.0902757Z d80602abf3cc: Waiting 2025-09-07T07:37:07.0902898Z 744523d9b7f5: Waiting 2025-09-07T07:37:07.0903220Z 4e2848642431: Waiting 2025-09-07T07:37:07.0903365Z e7769b0d7a82: Waiting 2025-09-07T07:37:07.0903512Z 7e9ff0c6f103: Waiting 2025-09-07T07:37:07.0903647Z 092516f71fe3: Waiting 2025-09-07T07:37:07.0903792Z cc93cd65e90f: Waiting 2025-09-07T07:37:07.0903938Z 7ace90c063f3: Waiting 2025-09-07T07:37:07.0904083Z 68c20f3c23bb: Waiting 2025-09-07T07:37:07.0904229Z 7ddca6c4c050: Waiting 2025-09-07T07:37:07.0904372Z 7efa39950d32: Waiting 2025-09-07T07:37:07.0904522Z 119ab3bceafa: Waiting 2025-09-07T07:37:07.0904669Z 4f4fb700ef54: Waiting 2025-09-07T07:37:07.0904993Z 5e944f1ed1be: Waiting 2025-09-07T07:37:07.0905144Z 80b2232d952f: Waiting 2025-09-07T07:37:07.0905298Z 8085756b0cc0: Waiting 2025-09-07T07:37:07.0905448Z 622d8cfb39ea: Waiting 2025-09-07T07:37:07.0905590Z a625cbbc05b9: Waiting 2025-09-07T07:37:07.0905733Z 284119a92cb1: Waiting 2025-09-07T07:37:07.0905872Z be86f8c4f654: Waiting 2025-09-07T07:37:07.0906018Z af8eadc9eaab: Waiting 2025-09-07T07:37:07.0906161Z 37e2336101eb: Waiting 2025-09-07T07:37:07.0906322Z f8690caa3ac5: Waiting 2025-09-07T07:37:07.0906472Z 386b0c49c498: Waiting 2025-09-07T07:37:07.0906616Z 41619248f604: Waiting 2025-09-07T07:37:07.0906754Z f1ac881fde33: Waiting 2025-09-07T07:37:07.0906899Z 96695940d842: Waiting 2025-09-07T07:37:07.0907045Z da8d8b696333: Waiting 2025-09-07T07:37:07.0907189Z 71f708151a84: Waiting 2025-09-07T07:37:07.0907342Z 2908d6baaa6b: Waiting 2025-09-07T07:37:07.0907503Z 43b14c67347e: Waiting 2025-09-07T07:37:07.0907653Z 6e726ef07b5d: Waiting 2025-09-07T07:37:07.0907794Z 04c04be7408f: Waiting 2025-09-07T07:37:07.0907940Z 364070434a64: Waiting 2025-09-07T07:37:07.1608606Z 18a5ee5b0e2e: Download complete 2025-09-07T07:37:07.2489477Z 1c35b7d4b67c: Verifying Checksum 2025-09-07T07:37:07.2489745Z 1c35b7d4b67c: Download complete 2025-09-07T07:37:07.3473779Z 68c20f3c23bb: Verifying Checksum 2025-09-07T07:37:07.3474010Z 68c20f3c23bb: Download complete 2025-09-07T07:37:07.4344906Z e6fdc8487bfe: Verifying Checksum 2025-09-07T07:37:07.4345166Z e6fdc8487bfe: Download complete 2025-09-07T07:37:07.4394786Z 7efa39950d32: Verifying Checksum 2025-09-07T07:37:07.4395041Z 7efa39950d32: Download complete 2025-09-07T07:37:07.5096948Z a10eb16a7271: Verifying Checksum 2025-09-07T07:37:07.5097230Z a10eb16a7271: Download complete 2025-09-07T07:37:07.5188741Z 7d52cf579654: Verifying Checksum 2025-09-07T07:37:07.5188943Z 7d52cf579654: Download complete 2025-09-07T07:37:07.6123062Z 46fb6a8b3e1d: Verifying Checksum 2025-09-07T07:37:07.6123343Z 46fb6a8b3e1d: Download complete 2025-09-07T07:37:07.6887444Z 5ad6977cc38e: Download complete 2025-09-07T07:37:07.7743006Z da63046995a2: Verifying Checksum 2025-09-07T07:37:07.7743293Z da63046995a2: Download complete 2025-09-07T07:37:07.8660683Z 78243fdb9906: Verifying Checksum 2025-09-07T07:37:07.8660925Z 78243fdb9906: Download complete 2025-09-07T07:37:08.1079904Z e6fdc8487bfe: Pull complete 2025-09-07T07:37:08.1183627Z 18a5ee5b0e2e: Pull complete 2025-09-07T07:37:08.6747773Z cb6a20fcf4e2: Verifying Checksum 2025-09-07T07:37:08.6748018Z cb6a20fcf4e2: Download complete 2025-09-07T07:37:08.6814298Z 4f4fb700ef54: Verifying Checksum 2025-09-07T07:37:08.6814479Z 4f4fb700ef54: Download complete 2025-09-07T07:37:08.7766645Z 69715d3ad3c4: Verifying Checksum 2025-09-07T07:37:08.7766886Z 69715d3ad3c4: Download complete 2025-09-07T07:37:08.8655224Z 7ace90c063f3: Verifying Checksum 2025-09-07T07:37:08.8655475Z 7ace90c063f3: Download complete 2025-09-07T07:37:08.9641292Z acbd5447dd14: Verifying Checksum 2025-09-07T07:37:08.9641512Z acbd5447dd14: Download complete 2025-09-07T07:37:09.0606754Z 744523d9b7f5: Verifying Checksum 2025-09-07T07:37:09.0607005Z 744523d9b7f5: Download complete 2025-09-07T07:37:09.1545073Z 5bd615a7b945: Verifying Checksum 2025-09-07T07:37:09.1545341Z 5bd615a7b945: Download complete 2025-09-07T07:37:09.2244894Z f4986a00e3ae: Verifying Checksum 2025-09-07T07:37:09.2245130Z f4986a00e3ae: Download complete 2025-09-07T07:37:09.3196233Z 21902f6e4f8c: Verifying Checksum 2025-09-07T07:37:09.3196455Z 21902f6e4f8c: Download complete 2025-09-07T07:37:09.3953002Z d80602abf3cc: Verifying Checksum 2025-09-07T07:37:09.3953252Z d80602abf3cc: Download complete 2025-09-07T07:37:09.4842869Z 3c51bf0bc362: Download complete 2025-09-07T07:37:09.5500360Z 119ab3bceafa: Verifying Checksum 2025-09-07T07:37:09.5500577Z 119ab3bceafa: Download complete 2025-09-07T07:37:09.6263797Z af8eadc9eaab: Verifying Checksum 2025-09-07T07:37:09.6264033Z af8eadc9eaab: Download complete 2025-09-07T07:37:09.7000500Z e7769b0d7a82: Verifying Checksum 2025-09-07T07:37:09.7000726Z e7769b0d7a82: Download complete 2025-09-07T07:37:10.2820055Z 572424b92528: Verifying Checksum 2025-09-07T07:37:10.2820548Z 572424b92528: Download complete 2025-09-07T07:37:10.3639749Z a5ab7a280382: Verifying Checksum 2025-09-07T07:37:10.3639982Z a5ab7a280382: Download complete 2025-09-07T07:37:10.4466715Z 80b2232d952f: Verifying Checksum 2025-09-07T07:37:10.4466971Z 80b2232d952f: Download complete 2025-09-07T07:37:10.5384570Z cc93cd65e90f: Download complete 2025-09-07T07:37:10.6188400Z 0eed4c15712b: Verifying Checksum 2025-09-07T07:37:10.6188674Z 0eed4c15712b: Download complete 2025-09-07T07:37:10.8749679Z 092516f71fe3: Verifying Checksum 2025-09-07T07:37:10.8749915Z 092516f71fe3: Download complete 2025-09-07T07:37:10.9639085Z 8c0825014a62: Verifying Checksum 2025-09-07T07:37:10.9639325Z 8c0825014a62: Download complete 2025-09-07T07:37:11.0563651Z 8e0d2f63da0a: Verifying Checksum 2025-09-07T07:37:11.0563941Z 8e0d2f63da0a: Download complete 2025-09-07T07:37:11.1403893Z 73aae7958ba1: Verifying Checksum 2025-09-07T07:37:11.1404156Z 73aae7958ba1: Download complete 2025-09-07T07:37:11.2000474Z ac6077ec9fa5: Verifying Checksum 2025-09-07T07:37:11.2000722Z ac6077ec9fa5: Download complete 2025-09-07T07:37:11.2952614Z bf4ee4e45e92: Verifying Checksum 2025-09-07T07:37:11.2952885Z bf4ee4e45e92: Download complete 2025-09-07T07:37:14.3021466Z ba263639b0f4: Verifying Checksum 2025-09-07T07:37:14.3021749Z ba263639b0f4: Download complete 2025-09-07T07:37:14.3944367Z 6e726ef07b5d: Verifying Checksum 2025-09-07T07:37:14.3944609Z 6e726ef07b5d: Download complete 2025-09-07T07:37:16.5148185Z 572424b92528: Pull complete 2025-09-07T07:37:16.9136786Z 1c35b7d4b67c: Pull complete 2025-09-07T07:37:17.2051075Z 364070434a64: Verifying Checksum 2025-09-07T07:37:17.2051305Z 364070434a64: Download complete 2025-09-07T07:37:17.2652079Z 68c20f3c23bb: Pull complete 2025-09-07T07:37:17.6810200Z 7efa39950d32: Pull complete 2025-09-07T07:37:18.2043413Z a10eb16a7271: Pull complete 2025-09-07T07:37:18.7471748Z 7d52cf579654: Pull complete 2025-09-07T07:37:20.8591432Z cb6a20fcf4e2: Pull complete 2025-09-07T07:37:21.2300174Z 46fb6a8b3e1d: Pull complete 2025-09-07T07:37:21.6408070Z 5ad6977cc38e: Pull complete 2025-09-07T07:37:22.0704605Z da63046995a2: Pull complete 2025-09-07T07:37:22.5506864Z 78243fdb9906: Pull complete 2025-09-07T07:37:41.8836535Z 6f70d5d50aba: Verifying Checksum 2025-09-07T07:37:41.8836814Z 6f70d5d50aba: Download complete 2025-09-07T07:37:41.9656302Z 622d8cfb39ea: Verifying Checksum 2025-09-07T07:37:41.9656596Z 622d8cfb39ea: Download complete 2025-09-07T07:37:42.0330911Z 284119a92cb1: Verifying Checksum 2025-09-07T07:37:42.0331133Z 284119a92cb1: Download complete 2025-09-07T07:37:42.1132229Z 96695940d842: Verifying Checksum 2025-09-07T07:37:42.1132451Z 96695940d842: Download complete 2025-09-07T07:37:42.1769849Z 7ddca6c4c050: Verifying Checksum 2025-09-07T07:37:42.1770062Z 7ddca6c4c050: Download complete 2025-09-07T07:37:42.3018219Z a95e1f2f1aad: Verifying Checksum 2025-09-07T07:37:42.3018447Z a95e1f2f1aad: Download complete 2025-09-07T07:37:42.3944465Z 8085756b0cc0: Verifying Checksum 2025-09-07T07:37:42.3944703Z 8085756b0cc0: Download complete 2025-09-07T07:37:42.4736745Z 7e9ff0c6f103: Verifying Checksum 2025-09-07T07:37:42.4736980Z 7e9ff0c6f103: Download complete 2025-09-07T07:37:42.5403879Z a625cbbc05b9: Verifying Checksum 2025-09-07T07:37:42.5404105Z a625cbbc05b9: Download complete 2025-09-07T07:37:42.6130498Z 4e2848642431: Verifying Checksum 2025-09-07T07:37:42.6130741Z 4e2848642431: Download complete 2025-09-07T07:37:42.6931818Z 5e944f1ed1be: Verifying Checksum 2025-09-07T07:37:42.6932098Z 5e944f1ed1be: Download complete 2025-09-07T07:37:42.7789939Z 41619248f604: Verifying Checksum 2025-09-07T07:37:42.7790163Z 41619248f604: Download complete 2025-09-07T07:37:42.8799479Z be86f8c4f654: Verifying Checksum 2025-09-07T07:37:42.8799720Z be86f8c4f654: Download complete 2025-09-07T07:37:42.9683349Z ef1340e22a4b: Verifying Checksum 2025-09-07T07:37:42.9683624Z ef1340e22a4b: Download complete 2025-09-07T07:37:43.0617048Z da8d8b696333: Download complete 2025-09-07T07:37:45.5395487Z 386b0c49c498: Verifying Checksum 2025-09-07T07:37:45.5396265Z 386b0c49c498: Download complete 2025-09-07T07:37:45.6249265Z 2b1d0ea7efe0: Verifying Checksum 2025-09-07T07:37:45.6249521Z 2b1d0ea7efe0: Download complete 2025-09-07T07:37:45.6927621Z 04c04be7408f: Download complete 2025-09-07T07:37:45.7666809Z f8690caa3ac5: Download complete 2025-09-07T07:37:45.8330740Z 2908d6baaa6b: Verifying Checksum 2025-09-07T07:37:45.8330965Z 2908d6baaa6b: Download complete 2025-09-07T07:37:45.9792002Z 37e2336101eb: Verifying Checksum 2025-09-07T07:37:45.9792263Z 37e2336101eb: Download complete 2025-09-07T07:37:46.0585158Z f1ac881fde33: Download complete 2025-09-07T07:37:46.6681779Z 43b14c67347e: Verifying Checksum 2025-09-07T07:37:46.6682034Z 43b14c67347e: Download complete 2025-09-07T07:38:18.2569295Z 6f70d5d50aba: Pull complete 2025-09-07T07:38:18.7654563Z 4f4fb700ef54: Pull complete 2025-09-07T07:38:19.2038954Z 69715d3ad3c4: Pull complete 2025-09-07T07:38:19.7085737Z 7ace90c063f3: Pull complete 2025-09-07T07:38:20.1512568Z acbd5447dd14: Pull complete 2025-09-07T07:38:20.5793927Z 744523d9b7f5: Pull complete 2025-09-07T07:38:21.0230132Z 5bd615a7b945: Pull complete 2025-09-07T07:38:21.3319535Z f4986a00e3ae: Pull complete 2025-09-07T07:38:21.4092806Z 21902f6e4f8c: Pull complete 2025-09-07T07:38:21.5116558Z d80602abf3cc: Pull complete 2025-09-07T07:38:21.7974137Z 3c51bf0bc362: Pull complete 2025-09-07T07:38:21.8785484Z 71f708151a84: Verifying Checksum 2025-09-07T07:38:21.8785731Z 71f708151a84: Download complete 2025-09-07T07:38:22.2262062Z 119ab3bceafa: Pull complete 2025-09-07T07:38:22.9059793Z af8eadc9eaab: Pull complete 2025-09-07T07:38:23.2412022Z e7769b0d7a82: Pull complete 2025-09-07T07:38:28.7108976Z ba263639b0f4: Pull complete 2025-09-07T07:38:28.9369406Z a5ab7a280382: Pull complete 2025-09-07T07:38:29.2910877Z 80b2232d952f: Pull complete 2025-09-07T07:38:30.2807257Z cc93cd65e90f: Pull complete 2025-09-07T07:38:30.7226098Z 0eed4c15712b: Pull complete 2025-09-07T07:38:31.3945911Z 092516f71fe3: Pull complete 2025-09-07T07:38:31.9311261Z 8c0825014a62: Pull complete 2025-09-07T07:38:32.4647867Z 8e0d2f63da0a: Pull complete 2025-09-07T07:38:33.5231979Z 73aae7958ba1: Pull complete 2025-09-07T07:38:34.0487012Z ac6077ec9fa5: Pull complete 2025-09-07T07:38:34.5842570Z bf4ee4e45e92: Pull complete 2025-09-07T07:40:17.5103539Z c1b766f9b961: Verifying Checksum 2025-09-07T07:40:17.5103790Z c1b766f9b961: Download complete 2025-09-07T07:42:05.4316781Z c1b766f9b961: Pull complete 2025-09-07T07:42:05.9032320Z 6e726ef07b5d: Pull complete 2025-09-07T07:42:07.1109072Z 364070434a64: Pull complete 2025-09-07T07:43:14.3805641Z 71f708151a84: Pull complete 2025-09-07T07:43:14.8246549Z 622d8cfb39ea: Pull complete 2025-09-07T07:43:15.2830246Z 284119a92cb1: Pull complete 2025-09-07T07:43:16.2299932Z 96695940d842: Pull complete 2025-09-07T07:43:17.2300658Z 7ddca6c4c050: Pull complete 2025-09-07T07:43:17.6875382Z a95e1f2f1aad: Pull complete 2025-09-07T07:43:18.6914859Z 8085756b0cc0: Pull complete 2025-09-07T07:43:19.7368229Z 7e9ff0c6f103: Pull complete 2025-09-07T07:43:20.1650045Z a625cbbc05b9: Pull complete 2025-09-07T07:43:20.9255587Z 4e2848642431: Pull complete 2025-09-07T07:43:21.2464870Z 5e944f1ed1be: Pull complete 2025-09-07T07:43:22.2106296Z 41619248f604: Pull complete 2025-09-07T07:43:22.6109513Z be86f8c4f654: Pull complete 2025-09-07T07:43:23.3287122Z ef1340e22a4b: Pull complete 2025-09-07T07:43:23.8259615Z da8d8b696333: Pull complete 2025-09-07T07:43:29.1804076Z 386b0c49c498: Pull complete 2025-09-07T07:43:29.7143003Z 2b1d0ea7efe0: Pull complete 2025-09-07T07:43:30.1170180Z 04c04be7408f: Pull complete 2025-09-07T07:43:30.5475300Z f8690caa3ac5: Pull complete 2025-09-07T07:43:31.0887855Z 2908d6baaa6b: Pull complete 2025-09-07T07:43:31.4882558Z 37e2336101eb: Pull complete 2025-09-07T07:43:32.4272942Z f1ac881fde33: Pull complete 2025-09-07T07:43:34.0058608Z 43b14c67347e: Pull complete 2025-09-07T07:43:34.4850591Z Digest: sha256:383efb45082f20b8c808cb0ba4df693a01359592233f641f1f486911ac320a9a 2025-09-07T07:43:34.5259635Z Status: Downloaded newer image for 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:43:34.5485140Z 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:43:34.5546710Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:43:34.5547330Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-09-07T07:43:34.5557566Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:43:34.5557800Z env: 2025-09-07T07:43:34.5557954Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:34.5558128Z ##[endgroup] 2025-09-07T07:43:34.5642775Z Prepare all required actions 2025-09-07T07:43:34.5696033Z ##[group]Run ./.github/actions/get-workflow-job-id 2025-09-07T07:43:34.5696255Z with: 2025-09-07T07:43:34.5696806Z github-token: *** 2025-09-07T07:43:34.5696972Z env: 2025-09-07T07:43:34.5697124Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:34.5697289Z ##[endgroup] 2025-09-07T07:43:34.5795013Z ##[group]Run set -eux 2025-09-07T07:43:34.5795200Z set -eux 2025-09-07T07:43:34.5795478Z python3 .github/scripts/get_workflow_job_id.py "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-09-07T07:43:34.5803956Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:43:34.5804195Z env: 2025-09-07T07:43:34.5804344Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:34.5804660Z GITHUB_TOKEN: *** 2025-09-07T07:43:34.5804824Z ##[endgroup] 2025-09-07T07:43:34.5831864Z + python3 .github/scripts/get_workflow_job_id.py 17525294857 i-0d9447236daff7b96 2025-09-07T07:43:35.0038701Z Setting output job-id=49775530527 2025-09-07T07:43:35.0039257Z Setting output job-name=inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd) 2025-09-07T07:43:35.0167142Z ##[group]Run python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-09-07T07:43:35.0167632Z python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-09-07T07:43:35.0168208Z python3 -m tools.stats.monitor --log-interval "$MONITOR_LOG_INTERVAL" --data-collect-interval "$MONITOR_DATA_COLLECT_INTERVAL" > usage_log.txt 2>&1 & 2025-09-07T07:43:35.0168714Z echo "monitor-script-pid=${!}" >> "${GITHUB_OUTPUT}" 2025-09-07T07:43:35.0176784Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:43:35.0177016Z env: 2025-09-07T07:43:35.0177174Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:35.0177352Z JOB_ID: 49775530527 2025-09-07T07:43:35.0177745Z JOB_NAME: inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd) 2025-09-07T07:43:35.0178114Z WORKFLOW_NAME: inductor-perf-nightly-x86-zen 2025-09-07T07:43:35.0178371Z WORKFLOW_RUN_ID: 17525294857 2025-09-07T07:43:35.0178561Z MONITOR_LOG_INTERVAL: 15 2025-09-07T07:43:35.0178730Z MONITOR_DATA_COLLECT_INTERVAL: 4 2025-09-07T07:43:35.0178921Z ##[endgroup] 2025-09-07T07:43:35.5205831Z Defaulting to user installation because normal site-packages is not writeable 2025-09-07T07:43:35.7730907Z Collecting psutil==5.9.8 2025-09-07T07:43:35.7878549Z Downloading psutil-5.9.8-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (288 kB) 2025-09-07T07:43:35.8580832Z Collecting dataclasses_json==0.6.7 2025-09-07T07:43:35.8606785Z Downloading dataclasses_json-0.6.7-py3-none-any.whl (28 kB) 2025-09-07T07:43:35.8869997Z Collecting nvidia-ml-py==11.525.84 2025-09-07T07:43:35.8894079Z Downloading nvidia_ml_py-11.525.84-py3-none-any.whl (34 kB) 2025-09-07T07:43:35.9223429Z Collecting typing-inspect<1,>=0.4.0 2025-09-07T07:43:35.9248686Z Downloading typing_inspect-0.9.0-py3-none-any.whl (8.8 kB) 2025-09-07T07:43:36.0009877Z Collecting marshmallow<4.0.0,>=3.18.0 2025-09-07T07:43:36.0073492Z Downloading marshmallow-3.26.1-py3-none-any.whl (50 kB) 2025-09-07T07:43:36.0608504Z Collecting packaging>=17.0 2025-09-07T07:43:36.0633984Z Downloading packaging-25.0-py3-none-any.whl (66 kB) 2025-09-07T07:43:36.1137231Z Collecting mypy-extensions>=0.3.0 2025-09-07T07:43:36.1162126Z Downloading mypy_extensions-1.1.0-py3-none-any.whl (5.0 kB) 2025-09-07T07:43:36.1811387Z Collecting typing-extensions>=3.7.4 2025-09-07T07:43:36.1845633Z Downloading typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2025-09-07T07:43:36.3872802Z Installing collected packages: typing-extensions, packaging, mypy-extensions, typing-inspect, marshmallow, psutil, nvidia-ml-py, dataclasses-json 2025-09-07T07:43:36.9833959Z Successfully installed dataclasses-json-0.6.7 marshmallow-3.26.1 mypy-extensions-1.1.0 nvidia-ml-py-11.525.84 packaging-25.0 psutil-5.9.8 typing-extensions-4.15.0 typing-inspect-0.9.0 2025-09-07T07:43:37.2008650Z Prepare all required actions 2025-09-07T07:43:37.2008949Z Getting action download info 2025-09-07T07:43:37.3592676Z Download action repository 'seemethere/download-artifact-s3@v4' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-09-07T07:43:38.1815857Z Download action repository 'actions/download-artifact@v4' (SHA:d3f86a106a0bac45b974a628896c90dbdf5c8093) 2025-09-07T07:43:40.6931438Z ##[group]Run ./.github/actions/download-build-artifacts 2025-09-07T07:43:40.6931687Z with: 2025-09-07T07:43:40.6931854Z name: linux-jammy-py3.9-gcc11-build 2025-09-07T07:43:40.6932053Z s3-bucket: gha-artifacts 2025-09-07T07:43:40.6932225Z env: 2025-09-07T07:43:40.6932363Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:40.6932525Z ##[endgroup] 2025-09-07T07:43:40.7077397Z ##[group]Run seemethere/download-artifact-s3@v4 2025-09-07T07:43:40.7077616Z with: 2025-09-07T07:43:40.7077778Z name: linux-jammy-py3.9-gcc11-build 2025-09-07T07:43:40.7077984Z s3-bucket: gha-artifacts 2025-09-07T07:43:40.7078188Z region: us-east-1 2025-09-07T07:43:40.7078333Z env: 2025-09-07T07:43:40.7078478Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:40.7078657Z ##[endgroup] 2025-09-07T07:43:41.4753343Z (node:57460) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-09-07T07:43:41.4753666Z 2025-09-07T07:43:41.4753803Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-09-07T07:43:41.4754153Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-09-07T07:43:41.4754510Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-09-07T07:43:43.0852062Z Found 1 objects with prefix pytorch/pytorch/17525294857/linux-jammy-py3.9-gcc11-build/ 2025-09-07T07:43:43.0852575Z Starting download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-09-07T07:43:47.6751259Z Finished download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-09-07T07:43:47.6755992Z Artifact download has finished successfully 2025-09-07T07:43:47.7026640Z ##[group]Run unzip -o artifacts.zip 2025-09-07T07:43:47.7026901Z unzip -o artifacts.zip 2025-09-07T07:43:47.7035101Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:43:47.7035340Z env: 2025-09-07T07:43:47.7035491Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:47.7035667Z ##[endgroup] 2025-09-07T07:43:47.7268995Z Archive: artifacts.zip 2025-09-07T07:43:47.7269974Z creating: dist/ 2025-09-07T07:43:48.7115641Z inflating: dist/torch-2.9.0a0+git93fb23d-cp39-cp39-linux_x86_64.whl 2025-09-07T07:43:48.7223602Z inflating: dist/.ninja_log 2025-09-07T07:43:48.7224189Z creating: build/custom_test_artifacts/ 2025-09-07T07:43:48.7225075Z creating: build/custom_test_artifacts/custom-op-build/ 2025-09-07T07:43:48.7225446Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/ 2025-09-07T07:43:48.7225810Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/pkgRedirects/ 2025-09-07T07:43:48.7228921Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeConfigureLog.yaml 2025-09-07T07:43:48.7229881Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/ 2025-09-07T07:43:48.7230281Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeSystem.cmake 2025-09-07T07:43:48.7230701Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdC/ 2025-09-07T07:43:48.7231114Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdC/tmp/ 2025-09-07T07:43:48.7233054Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdC/CMakeCCompilerId.c 2025-09-07T07:43:48.7234113Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdC/a.out 2025-09-07T07:43:48.7234815Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeCCompiler.cmake 2025-09-07T07:43:48.7235485Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdCXX/ 2025-09-07T07:43:48.7235917Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdCXX/tmp/ 2025-09-07T07:43:48.7238385Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-09-07T07:43:48.7239632Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdCXX/a.out 2025-09-07T07:43:48.7240473Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeCXXCompiler.cmake 2025-09-07T07:43:48.7241809Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_C.bin 2025-09-07T07:43:48.7243300Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_CXX.bin 2025-09-07T07:43:48.7243757Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeScratch/ 2025-09-07T07:43:48.7244184Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/cmake.check_cache 2025-09-07T07:43:48.7244585Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/ 2025-09-07T07:43:48.7245036Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.ts 2025-09-07T07:43:48.7245529Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.make 2025-09-07T07:43:48.7246001Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/depend.make 2025-09-07T07:43:48.7247178Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/link.txt 2025-09-07T07:43:48.7247748Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/cmake_clean.cmake 2025-09-07T07:43:48.7248246Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/build.make 2025-09-07T07:43:48.7248772Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/DependInfo.cmake 2025-09-07T07:43:48.7249250Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/flags.make 2025-09-07T07:43:48.7249741Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/progress.make 2025-09-07T07:43:48.7267123Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o.d 2025-09-07T07:43:48.7434883Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o 2025-09-07T07:43:48.7435334Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/ 2025-09-07T07:43:48.7435819Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.ts 2025-09-07T07:43:48.7436356Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.make 2025-09-07T07:43:48.7436884Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/depend.make 2025-09-07T07:43:48.7437352Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/link.txt 2025-09-07T07:43:48.7438312Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/cmake_clean.cmake 2025-09-07T07:43:48.7438804Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/build.make 2025-09-07T07:43:48.7439294Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/DependInfo.cmake 2025-09-07T07:43:48.7439778Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/flags.make 2025-09-07T07:43:48.7440269Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/progress.make 2025-09-07T07:43:48.7457904Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o.d 2025-09-07T07:43:48.7525206Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o 2025-09-07T07:43:48.7525752Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-09-07T07:43:48.7526253Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/TargetDirectories.txt 2025-09-07T07:43:48.7526707Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/progress.marks 2025-09-07T07:43:48.7527389Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile2 2025-09-07T07:43:48.7529012Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile.cmake 2025-09-07T07:43:48.7529442Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/InstallScripts.json 2025-09-07T07:43:48.7530831Z inflating: build/custom_test_artifacts/custom-op-build/CMakeCache.txt 2025-09-07T07:43:48.7531582Z inflating: build/custom_test_artifacts/custom-op-build/Makefile 2025-09-07T07:43:48.7532349Z inflating: build/custom_test_artifacts/custom-op-build/cmake_install.cmake 2025-09-07T07:43:48.7676766Z inflating: build/custom_test_artifacts/custom-op-build/libcustom_ops.so 2025-09-07T07:43:48.7723570Z inflating: build/custom_test_artifacts/custom-op-build/test_custom_ops 2025-09-07T07:43:48.7723895Z creating: build/custom_test_artifacts/jit-hook-build/ 2025-09-07T07:43:48.7724196Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/ 2025-09-07T07:43:48.7724556Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/pkgRedirects/ 2025-09-07T07:43:48.7727797Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeConfigureLog.yaml 2025-09-07T07:43:48.7728188Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/ 2025-09-07T07:43:48.7728647Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeSystem.cmake 2025-09-07T07:43:48.7729082Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdC/ 2025-09-07T07:43:48.7729484Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdC/tmp/ 2025-09-07T07:43:48.7731534Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdC/CMakeCCompilerId.c 2025-09-07T07:43:48.7732772Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdC/a.out 2025-09-07T07:43:48.7733456Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeCCompiler.cmake 2025-09-07T07:43:48.7733882Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdCXX/ 2025-09-07T07:43:48.7734296Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdCXX/tmp/ 2025-09-07T07:43:48.7736384Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-09-07T07:43:48.7737742Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdCXX/a.out 2025-09-07T07:43:48.7739141Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeCXXCompiler.cmake 2025-09-07T07:43:48.7740423Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_C.bin 2025-09-07T07:43:48.7742225Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_CXX.bin 2025-09-07T07:43:48.7742678Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeScratch/ 2025-09-07T07:43:48.7743066Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/cmake.check_cache 2025-09-07T07:43:48.7743465Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/ 2025-09-07T07:43:48.7744170Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.ts 2025-09-07T07:43:48.7744964Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.make 2025-09-07T07:43:48.7745468Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/depend.make 2025-09-07T07:43:48.7745943Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/link.txt 2025-09-07T07:43:48.7746416Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/cmake_clean.cmake 2025-09-07T07:43:48.7746895Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/build.make 2025-09-07T07:43:48.7747368Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/DependInfo.cmake 2025-09-07T07:43:48.7747853Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/flags.make 2025-09-07T07:43:48.7748466Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/progress.make 2025-09-07T07:43:48.7765975Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o.d 2025-09-07T07:43:48.7818869Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o 2025-09-07T07:43:48.7819501Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-09-07T07:43:48.7820052Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/TargetDirectories.txt 2025-09-07T07:43:48.7820631Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/progress.marks 2025-09-07T07:43:48.7821269Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile2 2025-09-07T07:43:48.7822466Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile.cmake 2025-09-07T07:43:48.7823017Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/InstallScripts.json 2025-09-07T07:43:48.7824327Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeCache.txt 2025-09-07T07:43:48.7825093Z inflating: build/custom_test_artifacts/jit-hook-build/Makefile 2025-09-07T07:43:48.7825671Z inflating: build/custom_test_artifacts/jit-hook-build/cmake_install.cmake 2025-09-07T07:43:48.7857962Z inflating: build/custom_test_artifacts/jit-hook-build/test_jit_hooks 2025-09-07T07:43:48.7858291Z creating: build/custom_test_artifacts/custom-backend-build/ 2025-09-07T07:43:48.7858628Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/ 2025-09-07T07:43:48.7859013Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/pkgRedirects/ 2025-09-07T07:43:48.7862212Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeConfigureLog.yaml 2025-09-07T07:43:48.7862650Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/ 2025-09-07T07:43:48.7863082Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeSystem.cmake 2025-09-07T07:43:48.7863559Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdC/ 2025-09-07T07:43:48.7864005Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdC/tmp/ 2025-09-07T07:43:48.7865866Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdC/CMakeCCompilerId.c 2025-09-07T07:43:48.7867103Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdC/a.out 2025-09-07T07:43:48.7867758Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeCCompiler.cmake 2025-09-07T07:43:48.7868225Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdCXX/ 2025-09-07T07:43:48.7868681Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdCXX/tmp/ 2025-09-07T07:43:48.7870586Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-09-07T07:43:48.7871971Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdCXX/a.out 2025-09-07T07:43:48.7872658Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeCXXCompiler.cmake 2025-09-07T07:43:48.7874043Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_C.bin 2025-09-07T07:43:48.7875535Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_CXX.bin 2025-09-07T07:43:48.7876026Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeScratch/ 2025-09-07T07:43:48.7876451Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/cmake.check_cache 2025-09-07T07:43:48.7876882Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/ 2025-09-07T07:43:48.7877363Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.ts 2025-09-07T07:43:48.7877926Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.make 2025-09-07T07:43:48.7878465Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/depend.make 2025-09-07T07:43:48.7878961Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/link.txt 2025-09-07T07:43:48.7879480Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/cmake_clean.cmake 2025-09-07T07:43:48.7880183Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/build.make 2025-09-07T07:43:48.7880812Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/DependInfo.cmake 2025-09-07T07:43:48.7881345Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/flags.make 2025-09-07T07:43:48.7882123Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/progress.make 2025-09-07T07:43:48.7885864Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o.d 2025-09-07T07:43:48.7986404Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o 2025-09-07T07:43:48.7986916Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/ 2025-09-07T07:43:48.7987432Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.ts 2025-09-07T07:43:48.7987998Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.make 2025-09-07T07:43:48.7988542Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/depend.make 2025-09-07T07:43:48.7989048Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/link.txt 2025-09-07T07:43:48.7989584Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/cmake_clean.cmake 2025-09-07T07:43:48.7990443Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/build.make 2025-09-07T07:43:48.7990979Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/DependInfo.cmake 2025-09-07T07:43:48.7991513Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/flags.make 2025-09-07T07:43:48.7992034Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/progress.make 2025-09-07T07:43:48.8009569Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o.d 2025-09-07T07:43:48.8054812Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o 2025-09-07T07:43:48.8055552Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-09-07T07:43:48.8056054Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/TargetDirectories.txt 2025-09-07T07:43:48.8056516Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/progress.marks 2025-09-07T07:43:48.8057296Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile2 2025-09-07T07:43:48.8058565Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile.cmake 2025-09-07T07:43:48.8059006Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/InstallScripts.json 2025-09-07T07:43:48.8060309Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeCache.txt 2025-09-07T07:43:48.8061083Z inflating: build/custom_test_artifacts/custom-backend-build/Makefile 2025-09-07T07:43:48.8061611Z inflating: build/custom_test_artifacts/custom-backend-build/cmake_install.cmake 2025-09-07T07:43:48.8146733Z inflating: build/custom_test_artifacts/custom-backend-build/libcustom_backend.so 2025-09-07T07:43:48.8179535Z inflating: build/custom_test_artifacts/custom-backend-build/test_custom_backend 2025-09-07T07:43:48.8179836Z creating: build/lib/ 2025-09-07T07:43:48.8249291Z inflating: build/lib/libprotobuf-lite.a 2025-09-07T07:43:48.8620820Z inflating: build/lib/libprotobuf.a 2025-09-07T07:43:48.9036325Z inflating: build/lib/libprotoc.a 2025-09-07T07:43:48.9044680Z inflating: build/lib/libpthreadpool.a 2025-09-07T07:43:48.9051614Z inflating: build/lib/libcpuinfo.a 2025-09-07T07:43:48.9058346Z inflating: build/lib/libcpuinfo_internals.a 2025-09-07T07:43:48.9059331Z inflating: build/lib/libclog.a 2025-09-07T07:43:48.9075254Z inflating: build/lib/libpytorch_qnnpack.a 2025-09-07T07:43:48.9077393Z inflating: build/lib/libnnpack_reference_layers.a 2025-09-07T07:43:48.9233141Z inflating: build/lib/libmicrokernels-prod.a 2025-09-07T07:43:48.9248263Z inflating: build/lib/libnnpack.a 2025-09-07T07:43:48.9984472Z inflating: build/lib/libmicrokernels-all.a 2025-09-07T07:43:49.0043601Z inflating: build/lib/libgtest.a 2025-09-07T07:43:49.0057885Z inflating: build/lib/libgmock.a 2025-09-07T07:43:49.0058610Z inflating: build/lib/libgtest_main.a 2025-09-07T07:43:49.0059524Z inflating: build/lib/libgmock_main.a 2025-09-07T07:43:49.0136127Z inflating: build/lib/libXNNPACK.a 2025-09-07T07:43:49.0200306Z inflating: build/lib/libbenchmark.a 2025-09-07T07:43:49.0201021Z inflating: build/lib/libbenchmark_main.a 2025-09-07T07:43:49.0201696Z inflating: build/lib/libjitprofiling.a 2025-09-07T07:43:49.0208938Z inflating: build/lib/libittnotify.a 2025-09-07T07:43:49.0264451Z inflating: build/lib/libasmjit.a 2025-09-07T07:43:49.1264928Z inflating: build/lib/libfbgemm.a 2025-09-07T07:43:49.1290807Z inflating: build/lib/libtensorpipe_uv.a 2025-09-07T07:43:49.1754365Z inflating: build/lib/libtensorpipe.a 2025-09-07T07:43:49.1855744Z inflating: build/lib/libgloo.a 2025-09-07T07:43:49.1895103Z inflating: build/lib/libonnx_proto.a 2025-09-07T07:43:49.2498151Z inflating: build/lib/libonnx.a 2025-09-07T07:43:50.1028902Z inflating: build/lib/libdnnl.a 2025-09-07T07:43:50.1044112Z inflating: build/lib/libfmt.a 2025-09-07T07:43:50.1262876Z inflating: build/lib/libkineto.a 2025-09-07T07:43:50.1356444Z inflating: build/lib/libc10.so 2025-09-07T07:43:50.1357742Z inflating: build/lib/libtorch_global_deps.so 2025-09-07T07:43:52.6387769Z inflating: build/lib/libtorch_cpu.so 2025-09-07T07:43:52.6388813Z inflating: build/lib/libtorch.so 2025-09-07T07:43:52.6448878Z inflating: build/lib/libtorchbind_test.so 2025-09-07T07:43:52.6464473Z inflating: build/lib/libjitbackend_test.so 2025-09-07T07:43:52.6484905Z inflating: build/lib/libbackend_with_compiler.so 2025-09-07T07:43:52.6507571Z inflating: build/lib/libaoti_custom_ops.so 2025-09-07T07:43:52.6511326Z inflating: build/lib/libshm.so 2025-09-07T07:43:52.8233799Z inflating: build/lib/libtorch_python.so 2025-09-07T07:43:52.8263830Z inflating: build/lib/libnnapi_backend.so 2025-09-07T07:43:52.8264690Z creating: build/bin/ 2025-09-07T07:43:52.8264941Z creating: build/bin/CMakeFiles/ 2025-09-07T07:43:52.8265987Z inflating: build/bin/cmake_install.cmake 2025-09-07T07:43:52.8266244Z inflating: build/bin/CTestTestfile.cmake 2025-09-07T07:43:52.8647957Z inflating: build/bin/protoc-3.13.0.0 2025-09-07T07:43:52.9029620Z inflating: build/bin/protoc 2025-09-07T07:43:52.9076410Z inflating: build/bin/c10_CompileTimeFunctionPointer_test 2025-09-07T07:43:52.9125832Z inflating: build/bin/c10_AllocatorConfig_test 2025-09-07T07:43:52.9174011Z inflating: build/bin/c10_DeviceGuard_test 2025-09-07T07:43:52.9222488Z inflating: build/bin/c10_Device_test 2025-09-07T07:43:52.9268786Z inflating: build/bin/c10_StreamGuard_test 2025-09-07T07:43:52.9323497Z inflating: build/bin/c10_DispatchKeySet_test 2025-09-07T07:43:52.9373753Z inflating: build/bin/c10_Scalar_test 2025-09-07T07:43:52.9426423Z inflating: build/bin/c10_SymInt_test 2025-09-07T07:43:52.9477198Z inflating: build/bin/c10_InlineDeviceGuard_test 2025-09-07T07:43:52.9528499Z inflating: build/bin/c10_InlineStreamGuard_test 2025-09-07T07:43:52.9580399Z inflating: build/bin/c10_SizesAndStrides_test 2025-09-07T07:43:52.9644386Z inflating: build/bin/c10_cow_test 2025-09-07T07:43:52.9690789Z inflating: build/bin/c10_ArrayRef_test 2025-09-07T07:43:52.9736379Z inflating: build/bin/c10_ConstexprCrc_test 2025-09-07T07:43:52.9785995Z inflating: build/bin/c10_Bitset_test 2025-09-07T07:43:52.9832305Z inflating: build/bin/c10_DeadlockDetection_test 2025-09-07T07:43:52.9884975Z inflating: build/bin/c10_Enumerate_test 2025-09-07T07:43:52.9932481Z inflating: build/bin/c10_Half_test 2025-09-07T07:43:52.9981383Z inflating: build/bin/c10_IntrusiveList_test 2025-09-07T07:43:53.0033140Z inflating: build/bin/c10_LeftRight_test 2025-09-07T07:43:53.0084356Z inflating: build/bin/c10_Metaprogramming_test 2025-09-07T07:43:53.0133826Z inflating: build/bin/c10_NetworkFlow_test 2025-09-07T07:43:53.0180291Z inflating: build/bin/c10_Synchronized_test 2025-09-07T07:43:53.0227070Z inflating: build/bin/c10_Semaphore_test 2025-09-07T07:43:53.0274972Z inflating: build/bin/c10_TypeIndex_test 2025-09-07T07:43:53.0322432Z inflating: build/bin/c10_TypeList_test 2025-09-07T07:43:53.0373631Z inflating: build/bin/c10_ThreadLocal_test 2025-09-07T07:43:53.0419831Z inflating: build/bin/c10_TypeTraits_test 2025-09-07T07:43:53.0467695Z inflating: build/bin/c10_accumulate_test 2025-09-07T07:43:53.0520032Z inflating: build/bin/c10_bfloat16_test 2025-09-07T07:43:53.0566899Z inflating: build/bin/c10_bit_cast_test 2025-09-07T07:43:53.0619629Z inflating: build/bin/c10_complex_math_test 2025-09-07T07:43:53.0670783Z inflating: build/bin/c10_complex_test 2025-09-07T07:43:53.0716915Z inflating: build/bin/c10_error_test 2025-09-07T07:43:53.0765926Z inflating: build/bin/c10_exception_test 2025-09-07T07:43:53.0812620Z inflating: build/bin/c10_flags_test 2025-09-07T07:43:53.0859940Z inflating: build/bin/c10_irange_test 2025-09-07T07:43:53.0906759Z inflating: build/bin/c10_generic_math_test 2025-09-07T07:43:53.1050901Z inflating: build/bin/c10_intrusive_ptr_test 2025-09-07T07:43:53.1100274Z inflating: build/bin/c10_lazy_test 2025-09-07T07:43:53.1153173Z inflating: build/bin/c10_logging_test 2025-09-07T07:43:53.1202314Z inflating: build/bin/c10_registry_test 2025-09-07T07:43:53.1271117Z inflating: build/bin/c10_optional_test 2025-09-07T07:43:53.1327607Z inflating: build/bin/c10_ordered_preserving_dict_test 2025-09-07T07:43:53.1463527Z inflating: build/bin/c10_small_vector_test 2025-09-07T07:43:53.1511842Z inflating: build/bin/c10_ssize_test 2025-09-07T07:43:53.1564669Z inflating: build/bin/c10_string_util_test 2025-09-07T07:43:53.1610097Z inflating: build/bin/c10_string_view_test 2025-09-07T07:43:53.1656816Z inflating: build/bin/c10_tempfile_test 2025-09-07T07:43:53.1697321Z inflating: build/bin/c10_intrusive_ptr_benchmark 2025-09-07T07:43:53.1749504Z inflating: build/bin/c10_typeid_test 2025-09-07T07:43:53.2254523Z inflating: build/bin/vec_test_all_types_DEFAULT 2025-09-07T07:43:53.2772609Z inflating: build/bin/vec_test_all_types_AVX512 2025-09-07T07:43:53.3297155Z inflating: build/bin/vec_test_all_types_AVX2 2025-09-07T07:43:53.3347018Z inflating: build/bin/static_runtime_bench 2025-09-07T07:43:53.3564752Z inflating: build/bin/static_runtime_test 2025-09-07T07:43:53.3632056Z inflating: build/bin/Dict_test 2025-09-07T07:43:53.3680458Z inflating: build/bin/Dimname_test 2025-09-07T07:43:53.3740189Z inflating: build/bin/MaybeOwned_test 2025-09-07T07:43:53.3792448Z inflating: build/bin/NamedTensor_test 2025-09-07T07:43:53.3847490Z inflating: build/bin/apply_utils_test 2025-09-07T07:43:53.3901954Z inflating: build/bin/atest 2025-09-07T07:43:53.3960598Z inflating: build/bin/basic 2025-09-07T07:43:53.4012285Z inflating: build/bin/broadcast_test 2025-09-07T07:43:53.4059463Z inflating: build/bin/cpu_allocator_test 2025-09-07T07:43:53.4113481Z inflating: build/bin/cpu_generator_test 2025-09-07T07:43:53.4162701Z inflating: build/bin/cpu_profiling_allocator_test 2025-09-07T07:43:53.4245444Z inflating: build/bin/cpu_rng_test 2025-09-07T07:43:53.4292748Z inflating: build/bin/dlconvertor_test 2025-09-07T07:43:53.4345501Z inflating: build/bin/extension_backend_test 2025-09-07T07:43:53.4396991Z inflating: build/bin/half_test 2025-09-07T07:43:53.4482955Z inflating: build/bin/ivalue_test 2025-09-07T07:43:53.4529479Z inflating: build/bin/lazy_tensor_test 2025-09-07T07:43:53.4578808Z inflating: build/bin/math_kernel_test 2025-09-07T07:43:53.4628484Z inflating: build/bin/memory_format_test 2025-09-07T07:43:53.4677919Z inflating: build/bin/memory_overlapping_test 2025-09-07T07:43:53.4727433Z inflating: build/bin/mobile_memory_cleanup 2025-09-07T07:43:53.4779209Z inflating: build/bin/native_test 2025-09-07T07:43:53.4827048Z inflating: build/bin/operator_name_test 2025-09-07T07:43:53.4874039Z inflating: build/bin/operators_test 2025-09-07T07:43:53.4922662Z inflating: build/bin/packedtensoraccessor_test 2025-09-07T07:43:53.4984063Z inflating: build/bin/pow_test 2025-09-07T07:43:53.5037021Z inflating: build/bin/quantized_test 2025-09-07T07:43:53.5083126Z inflating: build/bin/reduce_ops_test 2025-09-07T07:43:53.5130698Z inflating: build/bin/reportMemoryUsage_test 2025-09-07T07:43:53.5182366Z inflating: build/bin/scalar_tensor_test 2025-09-07T07:43:53.5236968Z inflating: build/bin/scalar_test 2025-09-07T07:43:53.5284158Z inflating: build/bin/StorageUtils_test 2025-09-07T07:43:53.5332891Z inflating: build/bin/stride_properties_test 2025-09-07T07:43:53.5404121Z inflating: build/bin/tensor_iterator_test 2025-09-07T07:43:53.5454358Z inflating: build/bin/test_parallel 2025-09-07T07:43:53.5501269Z inflating: build/bin/thread_init_test 2025-09-07T07:43:53.5552521Z inflating: build/bin/type_ptr_test 2025-09-07T07:43:53.5607029Z inflating: build/bin/type_test 2025-09-07T07:43:53.5656068Z inflating: build/bin/undefined_tensor_test 2025-09-07T07:43:53.5702141Z inflating: build/bin/verify_api_visibility 2025-09-07T07:43:53.5766314Z inflating: build/bin/legacy_vmap_test 2025-09-07T07:43:53.5814258Z inflating: build/bin/weakref_test 2025-09-07T07:43:53.5861723Z inflating: build/bin/wrapdim_test 2025-09-07T07:43:53.5909583Z inflating: build/bin/xla_tensor_test 2025-09-07T07:43:53.5964229Z inflating: build/bin/IListRef_test 2025-09-07T07:43:53.6058463Z inflating: build/bin/List_test 2025-09-07T07:43:53.6118937Z inflating: build/bin/KernelFunction_test 2025-09-07T07:43:53.6226208Z inflating: build/bin/kernel_function_legacy_test 2025-09-07T07:43:53.6310952Z inflating: build/bin/kernel_function_test 2025-09-07T07:43:53.6422428Z inflating: build/bin/kernel_lambda_legacy_test 2025-09-07T07:43:53.6513315Z inflating: build/bin/kernel_lambda_test 2025-09-07T07:43:53.6568700Z inflating: build/bin/kernel_stackbased_test 2025-09-07T07:43:53.6654066Z inflating: build/bin/make_boxed_from_unboxed_functor_test 2025-09-07T07:43:53.6701404Z inflating: build/bin/CppSignature_test 2025-09-07T07:43:53.6752433Z inflating: build/bin/backend_fallback_test 2025-09-07T07:43:53.6798186Z inflating: build/bin/op_allowlist_test 2025-09-07T07:43:53.7067781Z inflating: build/bin/op_registration_test 2025-09-07T07:43:53.7129026Z inflating: build/bin/inline_container_test 2025-09-07T07:43:53.8071628Z inflating: build/bin/test_jit 2025-09-07T07:43:53.8121101Z inflating: build/bin/FileStoreTest 2025-09-07T07:43:53.8169971Z inflating: build/bin/BackoffTest 2025-09-07T07:43:53.8222812Z inflating: build/bin/TCPStoreTest 2025-09-07T07:43:53.8545491Z inflating: build/bin/test_nativert 2025-09-07T07:43:53.8595362Z inflating: build/bin/HashStoreTest 2025-09-07T07:43:53.8655769Z inflating: build/bin/ProcessGroupGlooTest 2025-09-07T07:43:53.8658378Z inflating: build/bin/example_allreduce 2025-09-07T07:43:53.8709153Z inflating: build/bin/test_dist_autograd 2025-09-07T07:43:53.8770489Z inflating: build/bin/test_cpp_rpc 2025-09-07T07:43:53.9738642Z inflating: build/bin/test_api 2025-09-07T07:43:53.9740900Z inflating: build/bin/parallel_benchmark 2025-09-07T07:43:54.0037477Z inflating: build/bin/test_lazy 2025-09-07T07:43:54.0041241Z inflating: build/bin/torch_shm_manager 2025-09-07T07:43:54.0041514Z creating: .additional_ci_files/ 2025-09-07T07:43:54.0116668Z inflating: .additional_ci_files/test-times.json 2025-09-07T07:43:54.0401364Z inflating: .additional_ci_files/test-class-times.json 2025-09-07T07:43:54.0458858Z ##[group]Run rm artifacts.zip 2025-09-07T07:43:54.0459084Z rm artifacts.zip 2025-09-07T07:43:54.0467452Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:43:54.0467726Z env: 2025-09-07T07:43:54.0467876Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:54.0468054Z ##[endgroup] 2025-09-07T07:43:54.1605480Z ##[group]Run df -H 2025-09-07T07:43:54.1605668Z df -H 2025-09-07T07:43:54.1612860Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:43:54.1613093Z env: 2025-09-07T07:43:54.1613242Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:54.1613417Z ##[endgroup] 2025-09-07T07:43:54.1662460Z Filesystem Size Used Avail Use% Mounted on 2025-09-07T07:43:54.1662745Z devtmpfs 4.2M 0 4.2M 0% /dev 2025-09-07T07:43:54.1662957Z tmpfs 199G 0 199G 0% /dev/shm 2025-09-07T07:43:54.1663188Z tmpfs 80G 1.3M 80G 1% /run 2025-09-07T07:43:54.1663412Z /dev/nvme0n1p1 215G 72G 144G 34% / 2025-09-07T07:43:54.1663658Z tmpfs 199G 13k 199G 1% /tmp 2025-09-07T07:43:54.1663877Z /dev/nvme0n1p128 11M 1.4M 9.2M 13% /boot/efi 2025-09-07T07:43:54.1690969Z Prepare all required actions 2025-09-07T07:43:54.1691772Z Getting action download info 2025-09-07T07:43:54.2975574Z ##[group]Run ./.github/actions/download-td-artifacts 2025-09-07T07:43:54.2975815Z with: 2025-09-07T07:43:54.2975956Z env: 2025-09-07T07:43:54.2976098Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:54.2976263Z ##[endgroup] 2025-09-07T07:43:54.3018905Z ##[group]Run seemethere/download-artifact-s3@v4 2025-09-07T07:43:54.3019121Z with: 2025-09-07T07:43:54.3019253Z name: td_results 2025-09-07T07:43:54.3019410Z s3-bucket: gha-artifacts 2025-09-07T07:43:54.3019585Z region: us-east-1 2025-09-07T07:43:54.3019733Z env: 2025-09-07T07:43:54.3019870Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:54.3020044Z ##[endgroup] 2025-09-07T07:43:54.6641449Z (node:57483) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-09-07T07:43:54.6641771Z 2025-09-07T07:43:54.6641956Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-09-07T07:43:54.6642321Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-09-07T07:43:54.7758730Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-09-07T07:43:54.7759093Z Found 0 objects with prefix pytorch/pytorch/17525294857/td_results/ 2025-09-07T07:43:54.7763842Z Artifact download has finished successfully 2025-09-07T07:43:54.8048972Z ##[group]Run mkdir -p .additional_ci_files 2025-09-07T07:43:54.8049228Z mkdir -p .additional_ci_files 2025-09-07T07:43:54.8049513Z mv td_results.json .additional_ci_files/td_results.json || true 2025-09-07T07:43:54.8056768Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:43:54.8057017Z env: 2025-09-07T07:43:54.8057166Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:54.8057344Z ##[endgroup] 2025-09-07T07:43:54.8115333Z mv: cannot stat 'td_results.json': No such file or directory 2025-09-07T07:43:54.8169773Z ##[group]Run .github/scripts/parse_ref.py 2025-09-07T07:43:54.8170072Z .github/scripts/parse_ref.py 2025-09-07T07:43:54.8177778Z shell: /usr/bin/bash -e {0} 2025-09-07T07:43:54.8177983Z env: 2025-09-07T07:43:54.8178149Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:54.8178348Z ##[endgroup] 2025-09-07T07:43:54.8787564Z Setting output branch=main 2025-09-07T07:43:54.8873918Z Prepare all required actions 2025-09-07T07:43:54.8874202Z Getting action download info 2025-09-07T07:43:55.0067656Z ##[group]Run ./.github/actions/filter-test-configs 2025-09-07T07:43:55.0067887Z with: 2025-09-07T07:43:55.0068190Z github-token: *** 2025-09-07T07:43:55.0070624Z test-matrix: {"include": [{"config": "inductor_huggingface_perf_cpu_x86_zen", "shard": 1, "num_shards": 3, "runner": "linux.24xlarge.amd"}, {"config": "inductor_huggingface_perf_cpu_x86_zen", "shard": 2, "num_shards": 3, "runner": "linux.24xlarge.amd"}, {"config": "inductor_huggingface_perf_cpu_x86_zen", "shard": 3, "num_shards": 3, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 1, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 2, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 3, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 4, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 5, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 1, "num_shards": 4, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 2, "num_shards": 4, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 3, "num_shards": 4, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 4, "num_shards": 4, "runner": "linux.24xlarge.amd"}]} 2025-09-07T07:43:55.0073354Z job-name: inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd) 2025-09-07T07:43:55.0073863Z env: 2025-09-07T07:43:55.0074006Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:55.0074182Z ##[endgroup] 2025-09-07T07:43:55.0156955Z ##[group]Run nick-fields/retry@v3.0.0 2025-09-07T07:43:55.0157153Z with: 2025-09-07T07:43:55.0157290Z shell: bash 2025-09-07T07:43:55.0157440Z timeout_minutes: 10 2025-09-07T07:43:55.0157600Z max_attempts: 5 2025-09-07T07:43:55.0157765Z retry_wait_seconds: 30 2025-09-07T07:43:55.0158239Z command: set -eux # PyYAML 6.0 doesn't work with MacOS x86 anymore # This must run on Python-3.7 (AmazonLinux2) so can't use request=3.32.2 python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-09-07T07:43:55.0158723Z polling_interval_seconds: 1 2025-09-07T07:43:55.0158910Z warning_on_retry: true 2025-09-07T07:43:55.0159081Z continue_on_error: false 2025-09-07T07:43:55.0159247Z env: 2025-09-07T07:43:55.0159376Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:55.0159667Z GITHUB_TOKEN: *** 2025-09-07T07:43:55.0159834Z ##[endgroup] 2025-09-07T07:43:55.1659204Z + python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-09-07T07:43:55.3386952Z Defaulting to user installation because normal site-packages is not writeable 2025-09-07T07:43:55.5760451Z Collecting requests==2.27.1 2025-09-07T07:43:55.6035743Z Downloading requests-2.27.1-py2.py3-none-any.whl (63 kB) 2025-09-07T07:43:55.8390556Z Collecting pyyaml==6.0.2 2025-09-07T07:43:55.8471216Z Downloading PyYAML-6.0.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (737 kB) 2025-09-07T07:43:55.9571820Z Requirement already satisfied: urllib3<1.27,>=1.21.1 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (1.25.10) 2025-09-07T07:43:56.2272095Z Collecting charset-normalizer~=2.0.0 2025-09-07T07:43:56.2352964Z Downloading charset_normalizer-2.0.12-py3-none-any.whl (39 kB) 2025-09-07T07:43:56.3609668Z Collecting certifi>=2017.4.17 2025-09-07T07:43:56.3688753Z Downloading certifi-2025.8.3-py3-none-any.whl (161 kB) 2025-09-07T07:43:56.4180646Z Requirement already satisfied: idna<4,>=2.5 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (2.10) 2025-09-07T07:43:56.4727515Z Installing collected packages: charset-normalizer, certifi, requests, pyyaml 2025-09-07T07:43:56.8209474Z Successfully installed certifi-2025.8.3 charset-normalizer-2.0.12 pyyaml-6.0.2 requests-2.27.1 2025-09-07T07:43:57.0763536Z Command completed after 1 attempt(s). 2025-09-07T07:43:57.0824059Z ##[group]Run set -x 2025-09-07T07:43:57.0824254Z set -x 2025-09-07T07:43:57.0824401Z  2025-09-07T07:43:57.0824641Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-09-07T07:43:57.0824936Z # in runner workspace 2025-09-07T07:43:57.0825189Z python3 "${GITHUB_ACTION_PATH}/../../scripts/parse_ref.py" 2025-09-07T07:43:57.0832411Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:43:57.0832648Z env: 2025-09-07T07:43:57.0832791Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:57.0832966Z ##[endgroup] 2025-09-07T07:43:57.0860106Z + python3 /home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/filter-test-configs/../../scripts/parse_ref.py 2025-09-07T07:43:57.1008068Z Setting output branch=main 2025-09-07T07:43:57.1051482Z ##[group]Run echo "Workflow: ${GITHUB_WORKFLOW}" 2025-09-07T07:43:57.1051764Z echo "Workflow: ${GITHUB_WORKFLOW}" 2025-09-07T07:43:57.1051978Z echo "Job name: ${JOB_NAME}" 2025-09-07T07:43:57.1052160Z  2025-09-07T07:43:57.1052427Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-09-07T07:43:57.1052730Z # in runner workspace 2025-09-07T07:43:57.1052996Z python3 "${GITHUB_ACTION_PATH}/../../scripts/filter_test_configs.py" \ 2025-09-07T07:43:57.1053299Z  --workflow "${GITHUB_WORKFLOW}" \ 2025-09-07T07:43:57.1053508Z  --job-name "${JOB_NAME}" \ 2025-09-07T07:43:57.1055954Z  --test-matrix "{"include": [{"config": "inductor_huggingface_perf_cpu_x86_zen", "shard": 1, "num_shards": 3, "runner": "linux.24xlarge.amd"}, {"config": "inductor_huggingface_perf_cpu_x86_zen", "shard": 2, "num_shards": 3, "runner": "linux.24xlarge.amd"}, {"config": "inductor_huggingface_perf_cpu_x86_zen", "shard": 3, "num_shards": 3, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 1, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 2, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 3, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 4, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 5, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 1, "num_shards": 4, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 2, "num_shards": 4, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 3, "num_shards": 4, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 4, "num_shards": 4, "runner": "linux.24xlarge.amd"}]}" \ 2025-09-07T07:43:57.1058770Z  --selected-test-configs "" \ 2025-09-07T07:43:57.1058977Z  --pr-number "${PR_NUMBER}" \ 2025-09-07T07:43:57.1059167Z  --tag "${TAG}" \ 2025-09-07T07:43:57.1059352Z  --event-name "${EVENT_NAME}" \ 2025-09-07T07:43:57.1059558Z  --schedule "${SCHEDULE}" \ 2025-09-07T07:43:57.1059743Z  --branch "${HEAD_BRANCH}" 2025-09-07T07:43:57.1066820Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:43:57.1067060Z env: 2025-09-07T07:43:57.1067205Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:57.1067615Z GITHUB_TOKEN: *** 2025-09-07T07:43:57.1067932Z JOB_NAME: inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd) 2025-09-07T07:43:57.1068276Z PR_NUMBER: 2025-09-07T07:43:57.1068423Z TAG: 2025-09-07T07:43:57.1068561Z EVENT_NAME: schedule 2025-09-07T07:43:57.1068715Z SCHEDULE: 0 7 * * * 2025-09-07T07:43:57.1068874Z HEAD_BRANCH: main 2025-09-07T07:43:57.1069185Z ##[endgroup] 2025-09-07T07:43:57.1096230Z Workflow: inductor-perf-nightly-x86-zen 2025-09-07T07:43:57.1096623Z Job name: inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd) 2025-09-07T07:43:57.2585898Z Setting output keep-going=True 2025-09-07T07:43:57.2586172Z Setting output ci-verbose-test-logs=False 2025-09-07T07:43:57.2586408Z Setting output ci-test-showlocals=False 2025-09-07T07:43:57.2586624Z Setting output ci-no-test-timeout=False 2025-09-07T07:43:57.2586834Z Setting output ci-no-td=False 2025-09-07T07:43:57.2587026Z Setting output ci-td-distributed=False 2025-09-07T07:43:57.2587238Z Setting output is-unstable=False 2025-09-07T07:43:57.2587436Z Setting output reenabled-issues= 2025-09-07T07:43:57.2589959Z Setting output test-matrix={"include": [{"config": "inductor_huggingface_perf_cpu_x86_zen", "shard": 1, "num_shards": 3, "runner": "linux.24xlarge.amd"}, {"config": "inductor_huggingface_perf_cpu_x86_zen", "shard": 2, "num_shards": 3, "runner": "linux.24xlarge.amd"}, {"config": "inductor_huggingface_perf_cpu_x86_zen", "shard": 3, "num_shards": 3, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 1, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 2, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 3, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 4, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 5, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 1, "num_shards": 4, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 2, "num_shards": 4, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 3, "num_shards": 4, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 4, "num_shards": 4, "runner": "linux.24xlarge.amd"}]} 2025-09-07T07:43:57.2593019Z Setting output is-test-matrix-empty=False 2025-09-07T07:43:57.2705620Z ##[group]Run echo "Filtered matrix:" 2025-09-07T07:43:57.2705865Z echo "Filtered matrix:" 2025-09-07T07:43:57.2708274Z echo "{"include": [{"config": "inductor_huggingface_perf_cpu_x86_zen", "shard": 1, "num_shards": 3, "runner": "linux.24xlarge.amd"}, {"config": "inductor_huggingface_perf_cpu_x86_zen", "shard": 2, "num_shards": 3, "runner": "linux.24xlarge.amd"}, {"config": "inductor_huggingface_perf_cpu_x86_zen", "shard": 3, "num_shards": 3, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 1, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 2, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 3, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 4, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_timm_perf_cpu_x86_zen", "shard": 5, "num_shards": 5, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 1, "num_shards": 4, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 2, "num_shards": 4, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 3, "num_shards": 4, "runner": "linux.24xlarge.amd"}, {"config": "inductor_torchbench_perf_cpu_x86_zen", "shard": 4, "num_shards": 4, "runner": "linux.24xlarge.amd"}]}" 2025-09-07T07:43:57.2710867Z  2025-09-07T07:43:57.2711001Z echo 2025-09-07T07:43:57.2711187Z echo "Is the current job unstable? False" 2025-09-07T07:43:57.2711395Z  2025-09-07T07:43:57.2711526Z echo 2025-09-07T07:43:57.2711685Z echo "Is keep-going label set? True" 2025-09-07T07:43:57.2712081Z  2025-09-07T07:43:57.2712214Z echo 2025-09-07T07:43:57.2712363Z echo "Reenabled issues? " 2025-09-07T07:43:57.2719538Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:43:57.2719778Z env: 2025-09-07T07:43:57.2719928Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:57.2720095Z ##[endgroup] 2025-09-07T07:43:57.2746370Z Filtered matrix: 2025-09-07T07:43:57.2749375Z {include: [{config: inductor_huggingface_perf_cpu_x86_zen, shard: 1, num_shards: 3, runner: linux.24xlarge.amd}, {config: inductor_huggingface_perf_cpu_x86_zen, shard: 2, num_shards: 3, runner: linux.24xlarge.amd}, {config: inductor_huggingface_perf_cpu_x86_zen, shard: 3, num_shards: 3, runner: linux.24xlarge.amd}, {config: inductor_timm_perf_cpu_x86_zen, shard: 1, num_shards: 5, runner: linux.24xlarge.amd}, {config: inductor_timm_perf_cpu_x86_zen, shard: 2, num_shards: 5, runner: linux.24xlarge.amd}, {config: inductor_timm_perf_cpu_x86_zen, shard: 3, num_shards: 5, runner: linux.24xlarge.amd}, {config: inductor_timm_perf_cpu_x86_zen, shard: 4, num_shards: 5, runner: linux.24xlarge.amd}, {config: inductor_timm_perf_cpu_x86_zen, shard: 5, num_shards: 5, runner: linux.24xlarge.amd}, {config: inductor_torchbench_perf_cpu_x86_zen, shard: 1, num_shards: 4, runner: linux.24xlarge.amd}, {config: inductor_torchbench_perf_cpu_x86_zen, shard: 2, num_shards: 4, runner: linux.24xlarge.amd}, {config: inductor_torchbench_perf_cpu_x86_zen, shard: 3, num_shards: 4, runner: linux.24xlarge.amd}, {config: inductor_torchbench_perf_cpu_x86_zen, shard: 4, num_shards: 4, runner: linux.24xlarge.amd}]} 2025-09-07T07:43:57.2751933Z 2025-09-07T07:43:57.2752014Z Is the current job unstable? False 2025-09-07T07:43:57.2752158Z 2025-09-07T07:43:57.2752240Z Is keep-going label set? True 2025-09-07T07:43:57.2752642Z 2025-09-07T07:43:57.2752715Z Reenabled issues? 2025-09-07T07:43:57.2825595Z ##[group]Run echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-09-07T07:43:57.2825954Z echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-09-07T07:43:57.2833038Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:43:57.2833285Z env: 2025-09-07T07:43:57.2833439Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:57.2833619Z JOB_TIMEOUT: 720 2025-09-07T07:43:57.2833767Z ##[endgroup] 2025-09-07T07:43:57.2971933Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:43:57.2972282Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:43:57.2972558Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-09-07T07:43:57.2979724Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T07:43:57.2979969Z env: 2025-09-07T07:43:57.2980118Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:57.2980310Z ##[endgroup] 2025-09-07T07:43:57.3138168Z ##[group]Run set -x 2025-09-07T07:43:57.3138417Z set -x 2025-09-07T07:43:57.3138555Z  2025-09-07T07:43:57.3138725Z if [[ $TEST_CONFIG == 'multigpu' ]]; then 2025-09-07T07:43:57.3138980Z  TEST_COMMAND=.ci/pytorch/multigpu-test.sh 2025-09-07T07:43:57.3139236Z elif [[ $BUILD_ENVIRONMENT == *onnx* ]]; then 2025-09-07T07:43:57.3139454Z  TEST_COMMAND=.ci/onnx/test.sh 2025-09-07T07:43:57.3139643Z else 2025-09-07T07:43:57.3139821Z  TEST_COMMAND=.ci/pytorch/test.sh 2025-09-07T07:43:57.3140017Z fi 2025-09-07T07:43:57.3140153Z  2025-09-07T07:43:57.3140329Z # Leaving 1GB for the runner and other things 2025-09-07T07:43:57.3140686Z TOTAL_AVAILABLE_MEMORY_IN_GB=$(awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo) 2025-09-07T07:43:57.3141236Z # https://docs.docker.com/engine/containers/resource_constraints/#--memory-swap-details, the 3GB swap 2025-09-07T07:43:57.3141666Z # comes from https://github.com/pytorch/test-infra/pull/6058 2025-09-07T07:43:57.3141995Z TOTAL_MEMORY_WITH_SWAP=$(("${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}" + 3)) 2025-09-07T07:43:57.3142245Z  2025-09-07T07:43:57.3142420Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-09-07T07:43:57.3142628Z  SHM_OPTS= 2025-09-07T07:43:57.3142792Z  JENKINS_USER= 2025-09-07T07:43:57.3143018Z  # ensure that docker container cleanly exits in 12 hours 2025-09-07T07:43:57.3143306Z  # if for some reason cleanup action doesn't stop container 2025-09-07T07:43:57.3143549Z  # when job is cancelled 2025-09-07T07:43:57.3143740Z  DOCKER_SHELL_CMD="sleep 12h" 2025-09-07T07:43:57.3143924Z else 2025-09-07T07:43:57.3144089Z  SHM_OPTS="--shm-size=${SHM_SIZE}" 2025-09-07T07:43:57.3144295Z  JENKINS_USER="--user jenkins" 2025-09-07T07:43:57.3144493Z  DOCKER_SHELL_CMD= 2025-09-07T07:43:57.3144664Z fi 2025-09-07T07:43:57.3144798Z  2025-09-07T07:43:57.3145008Z # detached container should get cleaned up by teardown_ec2_linux 2025-09-07T07:43:57.3145321Z # TODO: Stop building test binaries as part of the build phase 2025-09-07T07:43:57.3145684Z # Used for GPU_FLAG, SHM_OPTS, JENKINS_USER and DOCKER_SHELL_CMD since that doesn't play nice 2025-09-07T07:43:57.3145998Z # shellcheck disable=SC2086,SC2090 2025-09-07T07:43:57.3146208Z container_name=$(docker run \ 2025-09-07T07:43:57.3146404Z  ${GPU_FLAG:-} \ 2025-09-07T07:43:57.3146596Z  ${SCCACHE_SERVER_PORT_DOCKER_FLAG:-} \ 2025-09-07T07:43:57.3146812Z  -e BUILD_ENVIRONMENT \ 2025-09-07T07:43:57.3146996Z  -e PR_NUMBER \ 2025-09-07T07:43:57.3147169Z  -e GITHUB_ACTIONS \ 2025-09-07T07:43:57.3147345Z  -e GITHUB_REPOSITORY \ 2025-09-07T07:43:57.3147692Z  -e GITHUB_WORKFLOW \ 2025-09-07T07:43:57.3147873Z  -e GITHUB_JOB \ 2025-09-07T07:43:57.3148056Z  -e GITHUB_RUN_ID \ 2025-09-07T07:43:57.3148224Z  -e GITHUB_RUN_NUMBER \ 2025-09-07T07:43:57.3148409Z  -e GITHUB_RUN_ATTEMPT \ 2025-09-07T07:43:57.3148599Z  -e JOB_ID \ 2025-09-07T07:43:57.3148799Z  -e JOB_NAME \ 2025-09-07T07:43:57.3148989Z  -e BASE_SHA \ 2025-09-07T07:43:57.3149176Z  -e BRANCH \ 2025-09-07T07:43:57.3149341Z  -e SHA1 \ 2025-09-07T07:43:57.3149519Z  -e AWS_DEFAULT_REGION \ 2025-09-07T07:43:57.3149714Z  -e IN_WHEEL_TEST \ 2025-09-07T07:43:57.3149913Z  -e SHARD_NUMBER \ 2025-09-07T07:43:57.3150093Z  -e TEST_CONFIG \ 2025-09-07T07:43:57.3150271Z  -e NUM_TEST_SHARDS \ 2025-09-07T07:43:57.3150450Z  -e REENABLED_ISSUES \ 2025-09-07T07:43:57.3150644Z  -e CONTINUE_THROUGH_ERROR \ 2025-09-07T07:43:57.3150960Z  -e VERBOSE_TEST_LOGS \ 2025-09-07T07:43:57.3151149Z  -e TEST_SHOWLOCALS \ 2025-09-07T07:43:57.3151321Z  -e NO_TEST_TIMEOUT \ 2025-09-07T07:43:57.3151494Z  -e NO_TD \ 2025-09-07T07:43:57.3151662Z  -e TD_DISTRIBUTED \ 2025-09-07T07:43:57.3151842Z  -e PR_LABELS \ 2025-09-07T07:43:57.3152034Z  -e MAX_JOBS="$(nproc --ignore=2)" \ 2025-09-07T07:43:57.3152242Z  -e SCCACHE_BUCKET \ 2025-09-07T07:43:57.3152415Z  -e SCCACHE_REGION \ 2025-09-07T07:43:57.3152584Z  -e XLA_CUDA \ 2025-09-07T07:43:57.3152758Z  -e XLA_CLANG_CACHE_S3_BUCKET_NAME \ 2025-09-07T07:43:57.3152979Z  -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK \ 2025-09-07T07:43:57.3153208Z  -e PYTORCH_TEST_RERUN_DISABLED_TESTS \ 2025-09-07T07:43:57.3153436Z  -e SKIP_SCCACHE_INITIALIZATION=1 \ 2025-09-07T07:43:57.3153647Z  -e HUGGING_FACE_HUB_TOKEN \ 2025-09-07T07:43:57.3153854Z  -e VLLM_TEST_HUGGING_FACE_TOKEN \ 2025-09-07T07:43:57.3154069Z  -e SCRIBE_GRAPHQL_ACCESS_TOKEN \ 2025-09-07T07:43:57.3154274Z  -e DASHBOARD_TAG \ 2025-09-07T07:43:57.3154457Z  -e ARTIFACTS_FILE_SUFFIX \ 2025-09-07T07:43:57.3154681Z  --memory="${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}g" \ 2025-09-07T07:43:57.3154941Z  --memory-swap="${TOTAL_MEMORY_WITH_SWAP}g" \ 2025-09-07T07:43:57.3155195Z  --env-file="/tmp/github_env_${GITHUB_RUN_ID}" \ 2025-09-07T07:43:57.3155432Z  --security-opt seccomp=unconfined \ 2025-09-07T07:43:57.3155647Z  --cap-add=SYS_PTRACE \ 2025-09-07T07:43:57.3155833Z  --ipc=host \ 2025-09-07T07:43:57.3168420Z  ${SHM_OPTS} \ 2025-09-07T07:43:57.3168610Z  --tty \ 2025-09-07T07:43:57.3168766Z  --detach \ 2025-09-07T07:43:57.3168938Z  --name="${container_name}" \ 2025-09-07T07:43:57.3169147Z  ${JENKINS_USER} \ 2025-09-07T07:43:57.3169392Z  -v "${GITHUB_WORKSPACE}:/var/lib/jenkins/workspace" \ 2025-09-07T07:43:57.3169644Z  -w /var/lib/jenkins/workspace \ 2025-09-07T07:43:57.3169844Z  "${DOCKER_IMAGE}" \ 2025-09-07T07:43:57.3170019Z  ${DOCKER_SHELL_CMD} 2025-09-07T07:43:57.3170183Z ) 2025-09-07T07:43:57.3170365Z # Propagate download.pytorch.org IP to container 2025-09-07T07:43:57.3170768Z grep download.pytorch.org /etc/hosts | docker exec -i "${container_name}" sudo bash -c "/bin/cat >> /etc/hosts" 2025-09-07T07:43:57.3171194Z echo "DOCKER_CONTAINER_ID=${container_name}" >> "${GITHUB_ENV}" 2025-09-07T07:43:57.3171445Z  2025-09-07T07:43:57.3171624Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-09-07T07:43:57.3171968Z  docker exec -t "${container_name}" sh -c "python3 -m pip install -r .ci/docker/requirements-ci.txt" 2025-09-07T07:43:57.3172273Z fi 2025-09-07T07:43:57.3172513Z  2025-09-07T07:43:57.3172813Z docker exec -t "${container_name}" sh -c "python3 -m pip install $(echo dist/*.whl)[opt-einsum] && ${TEST_COMMAND}" 2025-09-07T07:43:57.3180015Z shell: /usr/bin/bash -e {0} 2025-09-07T07:43:57.3180189Z env: 2025-09-07T07:43:57.3180335Z GIT_DEFAULT_BRANCH: main 2025-09-07T07:43:57.3180550Z BUILD_ENVIRONMENT: linux-jammy-py3.9-gcc11-build 2025-09-07T07:43:57.3180768Z PR_NUMBER: 2025-09-07T07:43:57.3180924Z GITHUB_REPOSITORY: pytorch/pytorch 2025-09-07T07:43:57.3181152Z GITHUB_WORKFLOW: inductor-perf-nightly-x86-zen 2025-09-07T07:43:57.3181358Z GITHUB_JOB: test 2025-09-07T07:43:57.3181509Z GITHUB_RUN_ID: 17525294857 2025-09-07T07:43:57.3181672Z GITHUB_RUN_NUMBER: 91 2025-09-07T07:43:57.3181831Z GITHUB_RUN_ATTEMPT: 1 2025-09-07T07:43:57.3181982Z JOB_ID: 49775530527 2025-09-07T07:43:57.3182298Z JOB_NAME: inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd) 2025-09-07T07:43:57.3182623Z BRANCH: main 2025-09-07T07:43:57.3182916Z SHA1: 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:43:57.3183158Z BASE_SHA: 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:43:57.3183402Z TEST_CONFIG: inductor_torchbench_perf_cpu_x86_zen 2025-09-07T07:43:57.3183604Z SHARD_NUMBER: 2 2025-09-07T07:43:57.3183752Z NUM_TEST_SHARDS: 4 2025-09-07T07:43:57.3183897Z REENABLED_ISSUES: 2025-09-07T07:43:57.3184056Z CONTINUE_THROUGH_ERROR: True 2025-09-07T07:43:57.3184231Z VERBOSE_TEST_LOGS: False 2025-09-07T07:43:57.3184418Z TEST_SHOWLOCALS: False 2025-09-07T07:43:57.3184594Z NO_TEST_TIMEOUT: False 2025-09-07T07:43:57.3184745Z NO_TD: False 2025-09-07T07:43:57.3184883Z TD_DISTRIBUTED: False 2025-09-07T07:43:57.3185081Z SCCACHE_BUCKET: ossci-compiler-cache-circleci-v2 2025-09-07T07:43:57.3185301Z SCCACHE_REGION: us-east-1 2025-09-07T07:43:57.3185467Z SHM_SIZE: 1g 2025-09-07T07:43:57.3185974Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:43:57.3186492Z XLA_CUDA: 2025-09-07T07:43:57.3186714Z XLA_CLANG_CACHE_S3_BUCKET_NAME: ossci-compiler-clang-cache-circleci-xla 2025-09-07T07:43:57.3186992Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK: 0 2025-09-07T07:43:57.3187191Z PYTORCH_TEST_RERUN_DISABLED_TESTS: 0 2025-09-07T07:43:57.3187592Z DASHBOARD_TAG: training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true 2025-09-07T07:43:57.3188163Z VLLM_TEST_HUGGING_FACE_TOKEN: *** 2025-09-07T07:43:57.3188422Z HUGGING_FACE_HUB_TOKEN: *** 2025-09-07T07:43:57.3188682Z SCRIBE_GRAPHQL_ACCESS_TOKEN: *** 2025-09-07T07:43:57.3189021Z ARTIFACTS_FILE_SUFFIX: test-inductor_torchbench_perf_cpu_x86_zen-2-4-linux.24xlarge.amd_49775530527 2025-09-07T07:43:57.3189343Z ##[endgroup] 2025-09-07T07:43:57.3214352Z + [[ inductor_torchbench_perf_cpu_x86_zen == \m\u\l\t\i\g\p\u ]] 2025-09-07T07:43:57.3214641Z + [[ linux-jammy-py3.9-gcc11-build == *onnx* ]] 2025-09-07T07:43:57.3214881Z + TEST_COMMAND=.ci/pytorch/test.sh 2025-09-07T07:43:57.3217985Z ++ awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo 2025-09-07T07:43:57.3239526Z + TOTAL_AVAILABLE_MEMORY_IN_GB='368.765 ' 2025-09-07T07:43:57.3239873Z + TOTAL_MEMORY_WITH_SWAP=371 2025-09-07T07:43:57.3240129Z + [[ linux-jammy-py3.9-gcc11-build == *\s\3\9\0\x* ]] 2025-09-07T07:43:57.3240358Z + SHM_OPTS=--shm-size=1g 2025-09-07T07:43:57.3240534Z + JENKINS_USER='--user jenkins' 2025-09-07T07:43:57.3240720Z + DOCKER_SHELL_CMD= 2025-09-07T07:43:57.3249090Z +++ nproc --ignore=2 2025-09-07T07:43:57.3681524Z ++ docker run -e BUILD_ENVIRONMENT -e PR_NUMBER -e GITHUB_ACTIONS -e GITHUB_REPOSITORY -e GITHUB_WORKFLOW -e GITHUB_JOB -e GITHUB_RUN_ID -e GITHUB_RUN_NUMBER -e GITHUB_RUN_ATTEMPT -e JOB_ID -e JOB_NAME -e BASE_SHA -e BRANCH -e SHA1 -e AWS_DEFAULT_REGION -e IN_WHEEL_TEST -e SHARD_NUMBER -e TEST_CONFIG -e NUM_TEST_SHARDS -e REENABLED_ISSUES -e CONTINUE_THROUGH_ERROR -e VERBOSE_TEST_LOGS -e TEST_SHOWLOCALS -e NO_TEST_TIMEOUT -e NO_TD -e TD_DISTRIBUTED -e PR_LABELS -e MAX_JOBS=94 -e SCCACHE_BUCKET -e SCCACHE_REGION -e XLA_CUDA -e XLA_CLANG_CACHE_S3_BUCKET_NAME -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK -e PYTORCH_TEST_RERUN_DISABLED_TESTS -e SKIP_SCCACHE_INITIALIZATION=1 -e HUGGING_FACE_HUB_TOKEN -e VLLM_TEST_HUGGING_FACE_TOKEN -e SCRIBE_GRAPHQL_ACCESS_TOKEN -e DASHBOARD_TAG -e ARTIFACTS_FILE_SUFFIX --memory=368g --memory-swap=371g --env-file=/tmp/github_env_17525294857 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --ipc=host --shm-size=1g --tty --detach --name= --user jenkins -v /home/ec2-user/actions-runner/_work/pytorch/pytorch:/var/lib/jenkins/workspace -w /var/lib/jenkins/workspace 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T07:46:45.6681589Z + container_name=de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T07:46:45.6686807Z + grep download.pytorch.org /etc/hosts 2025-09-07T07:46:45.6693537Z + docker exec -i de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac sudo bash -c '/bin/cat >> /etc/hosts' 2025-09-07T07:46:45.7899897Z + echo DOCKER_CONTAINER_ID=de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T07:46:45.7900793Z + [[ linux-jammy-py3.9-gcc11-build == *\s\3\9\0\x* ]] 2025-09-07T07:46:45.7908922Z ++ echo dist/torch-2.9.0a0+git93fb23d-cp39-cp39-linux_x86_64.whl 2025-09-07T07:46:45.7911229Z + docker exec -t de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac sh -c 'python3 -m pip install dist/torch-2.9.0a0+git93fb23d-cp39-cp39-linux_x86_64.whl[opt-einsum] && .ci/pytorch/test.sh' 2025-09-07T07:46:46.1455453Z Processing ./dist/torch-2.9.0a0+git93fb23d-cp39-cp39-linux_x86_64.whl (from torch==2.9.0a0+git93fb23d) 2025-09-07T07:46:46.3412438Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (3.19.1) 2025-09-07T07:46:46.3414174Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (4.15.0) 2025-09-07T07:46:46.3416881Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (1.13.3) 2025-09-07T07:46:46.3419882Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (2.8.8) 2025-09-07T07:46:46.3421995Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (3.1.6) 2025-09-07T07:46:46.3424743Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (2025.3.0) 2025-09-07T07:46:46.3433884Z Requirement already satisfied: opt-einsum>=3.3 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (3.3.0) 2025-09-07T07:46:46.3699183Z Requirement already satisfied: numpy>=1.7 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from opt-einsum>=3.3->torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (1.22.4) 2025-09-07T07:46:46.3712876Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from sympy>=1.13.3->torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (1.3.0) 2025-09-07T07:46:46.3741215Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from jinja2->torch==2.9.0a0+git93fb23d->torch==2.9.0a0+git93fb23d) (3.0.2) 2025-09-07T07:46:47.0723122Z Installing collected packages: torch 2025-09-07T07:46:54.1385645Z ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts. 2025-09-07T07:46:54.1386913Z dall-e 0.1 requires torchvision, which is not installed. 2025-09-07T07:46:54.1387203Z effdet 0.4.1 requires torchvision, which is not installed. 2025-09-07T07:46:54.1387554Z pytorch-labs-segment-anything-fast 0.2 requires torchao, which is not installed. 2025-09-07T07:46:54.1388008Z pytorch-labs-segment-anything-fast 0.2 requires torchvision>=0.17.0.dev20231026, which is not installed. 2025-09-07T07:46:54.1388482Z timm 1.0.14 requires torchvision, which is not installed. 2025-09-07T07:46:54.1388816Z Successfully installed torch-2.9.0a0+git93fb23d 2025-09-07T07:46:54.2175973Z + export TERM=vt100 2025-09-07T07:46:54.2176178Z + TERM=vt100 2025-09-07T07:46:54.2178831Z ++ dirname .ci/pytorch/test.sh 2025-09-07T07:46:54.2189684Z + source .ci/pytorch/common.sh 2025-09-07T07:46:54.2193434Z +++ dirname .ci/pytorch/common.sh 2025-09-07T07:46:54.2202713Z ++ source .ci/pytorch/common_utils.sh 2025-09-07T07:46:54.2204543Z +++ declare -f -t trap_add 2025-09-07T07:46:54.2208187Z ++ set -ex -o pipefail 2025-09-07T07:46:54.2208407Z ++ [[ linux-jammy-py3.9-gcc11-build == *rocm* ]] 2025-09-07T07:46:54.2208635Z ++ BUILD_TEST_LIBTORCH=0 2025-09-07T07:46:54.2212587Z ++ dirname .ci/pytorch/test.sh 2025-09-07T07:46:54.2222400Z + source .ci/pytorch/common-build.sh 2025-09-07T07:46:54.2223811Z ++ [[ linux-jammy-py3.9-gcc11-build != *win-* ]] 2025-09-07T07:46:54.2231204Z ++++ dirname .ci/pytorch/common-build.sh 2025-09-07T07:46:54.2240032Z +++ cd .ci/pytorch 2025-09-07T07:46:54.2240425Z +++ pwd -P 2025-09-07T07:46:54.2242899Z ++ script_dir=/var/lib/jenkins/workspace/.ci/pytorch 2025-09-07T07:46:54.2243213Z ++ [[ linux-jammy-py3.9-gcc11-build == *-pch* ]] 2025-09-07T07:46:54.2243428Z ++ which sccache 2025-09-07T07:46:54.2264731Z ++ [[ -z ossci-compiler-cache-circleci-v2 ]] 2025-09-07T07:46:54.2264951Z ++ sccache --stop-server 2025-09-07T07:46:54.2295815Z ++ true 2025-09-07T07:46:54.2295981Z ++ rm -f /var/lib/jenkins/sccache_error.log 2025-09-07T07:46:54.2309666Z ++ trap_add sccache_epilogue EXIT 2025-09-07T07:46:54.2309924Z ++ trap_add_cmd=sccache_epilogue 2025-09-07T07:46:54.2310118Z ++ shift 2025-09-07T07:46:54.2310275Z ++ for trap_add_name in "$@" 2025-09-07T07:46:54.2315748Z ++++ trap -p EXIT 2025-09-07T07:46:54.2318373Z +++ eval 'extract_trap_cmd ' 2025-09-07T07:46:54.2318565Z ++++ extract_trap_cmd 2025-09-07T07:46:54.2318727Z ++++ printf '%s\n' '' 2025-09-07T07:46:54.2318933Z +++ printf '%s\n' sccache_epilogue 2025-09-07T07:46:54.2320759Z ++ trap -- ' 2025-09-07T07:46:54.2320914Z sccache_epilogue' EXIT 2025-09-07T07:46:54.2321221Z ++ [[ -n 1 ]] 2025-09-07T07:46:54.2321508Z ++ echo 'Skipping sccache server initialization, setting environment variables' 2025-09-07T07:46:54.2321883Z Skipping sccache server initialization, setting environment variables 2025-09-07T07:46:54.2322168Z ++ export SCCACHE_IDLE_TIMEOUT=0 2025-09-07T07:46:54.2322360Z ++ SCCACHE_IDLE_TIMEOUT=0 2025-09-07T07:46:54.2323841Z ++ export SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-09-07T07:46:54.2324141Z ++ SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-09-07T07:46:54.2324414Z ++ export RUST_LOG=sccache::server=error 2025-09-07T07:46:54.2324625Z ++ RUST_LOG=sccache::server=error 2025-09-07T07:46:54.2324812Z ++ sccache --zero-stats 2025-09-07T07:46:54.4064855Z Statistics zeroed. 2025-09-07T07:46:54.4074254Z ++ which ccache 2025-09-07T07:46:54.4097743Z + [[ linux-jammy-py3.9-gcc11-build != *rocm* ]] 2025-09-07T07:46:54.4098003Z + [[ linux-jammy-py3.9-gcc11-build != *s390x* ]] 2025-09-07T07:46:54.4098234Z + [[ -d /var/lib/jenkins/workspace ]] 2025-09-07T07:46:54.4101605Z ++ stat -c %u /var/lib/jenkins/workspace 2025-09-07T07:46:54.4115977Z + WORKSPACE_ORIGINAL_OWNER_ID=1000 2025-09-07T07:46:54.4116187Z + trap_add cleanup_workspace EXIT 2025-09-07T07:46:54.4116384Z + trap_add_cmd=cleanup_workspace 2025-09-07T07:46:54.4116559Z + shift 2025-09-07T07:46:54.4116697Z + for trap_add_name in "$@" 2025-09-07T07:46:54.4126060Z +++ trap -p EXIT 2025-09-07T07:46:54.4127429Z ++ eval 'extract_trap_cmd trap -- '\'' 2025-09-07T07:46:54.4127976Z sccache_epilogue'\'' EXIT' 2025-09-07T07:46:54.4128149Z +++ extract_trap_cmd trap -- ' 2025-09-07T07:46:54.4128333Z sccache_epilogue' EXIT 2025-09-07T07:46:54.4128494Z +++ printf '%s\n' ' 2025-09-07T07:46:54.4128646Z sccache_epilogue' 2025-09-07T07:46:54.4128822Z ++ printf '%s\n' cleanup_workspace 2025-09-07T07:46:54.4129957Z + trap -- ' 2025-09-07T07:46:54.4130107Z sccache_epilogue 2025-09-07T07:46:54.4130261Z cleanup_workspace' EXIT 2025-09-07T07:46:54.4130457Z + sudo chown -R jenkins /var/lib/jenkins/workspace 2025-09-07T07:46:55.2222778Z + git config --global --add safe.directory /var/lib/jenkins/workspace 2025-09-07T07:46:55.2243490Z + echo 'Environment variables:' 2025-09-07T07:46:55.2243687Z Environment variables: 2025-09-07T07:46:55.2243860Z + env 2025-09-07T07:46:55.2254752Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-09-07T07:46:55.2255068Z CONTINUE_THROUGH_ERROR=True 2025-09-07T07:46:55.2255289Z BUILD_ENVIRONMENT=linux-jammy-py3.9-gcc11-build 2025-09-07T07:46:55.2256121Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-09-07T07:46:55.2256316Z HOSTNAME=de0caba1838c 2025-09-07T07:46:55.2256681Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_82e2c6f3-282a-484f-91e3-31f2b15ab45b 2025-09-07T07:46:55.2257055Z GITHUB_ACTION=__run_2 2025-09-07T07:46:55.2257232Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-09-07T07:46:55.2257425Z GITHUB_RUN_NUMBER=91 2025-09-07T07:46:55.2257611Z TEST_CONFIG=inductor_torchbench_perf_cpu_x86_zen 2025-09-07T07:46:55.2257886Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-09-07T07:46:55.2258099Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-09-07T07:46:55.2258302Z SCCACHE_IDLE_TIMEOUT=0 2025-09-07T07:46:55.2258552Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-09-07T07:46:55.2258744Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-09-07T07:46:55.2258938Z GITHUB_REF_TYPE=branch 2025-09-07T07:46:55.2259125Z BASE_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:46:55.2259330Z XLA_CUDA= 2025-09-07T07:46:55.2259484Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-09-07T07:46:55.2259775Z HUGGING_FACE_HUB_TOKEN=*** 2025-09-07T07:46:55.2260118Z *** 2025-09-07T07:46:55.2260267Z GITHUB_REPOSITORY_ID=65600975 2025-09-07T07:46:55.2260446Z GITHUB_ACTIONS=true 2025-09-07T07:46:55.2260642Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-09-07T07:46:55.2260890Z SHA1=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:46:55.2261122Z GITHUB_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:46:55.2261584Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/inductor-perf-test-nightly-x86-zen.yml@refs/heads/main 2025-09-07T07:46:55.2261951Z UCC_HOME=/usr 2025-09-07T07:46:55.2262093Z VERBOSE_TEST_LOGS=False 2025-09-07T07:46:55.2262266Z GITHUB_REF=refs/heads/main 2025-09-07T07:46:55.2262436Z SHARD_NUMBER=2 2025-09-07T07:46:55.2262588Z GITHUB_REF_PROTECTED=true 2025-09-07T07:46:55.2262750Z HOME=/var/lib/jenkins 2025-09-07T07:46:55.2262937Z GITHUB_API_URL=https://api.github.com 2025-09-07T07:46:55.2263167Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-09-07T07:46:55.2263357Z UCX_COMMIT= 2025-09-07T07:46:55.2263492Z USE_SYSTEM_NCCL=1 2025-09-07T07:46:55.2263642Z NUM_TEST_SHARDS=4 2025-09-07T07:46:55.2263786Z UCX_HOME=/usr 2025-09-07T07:46:55.2264136Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_82e2c6f3-282a-484f-91e3-31f2b15ab45b 2025-09-07T07:46:55.2264667Z JOB_NAME=inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd) 2025-09-07T07:46:55.2265178Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_82e2c6f3-282a-484f-91e3-31f2b15ab45b 2025-09-07T07:46:55.2265648Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-09-07T07:46:55.2265943Z GITHUB_EVENT_NAME=schedule 2025-09-07T07:46:55.2266334Z DASHBOARD_TAG=training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true 2025-09-07T07:46:55.2266740Z GITHUB_RUN_ID=17525294857 2025-09-07T07:46:55.2267094Z INSTALLED_OPENBLAS= 2025-09-07T07:46:55.2267466Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_82e2c6f3-282a-484f-91e3-31f2b15ab45b 2025-09-07T07:46:55.2267865Z GITHUB_ACTOR=pytorchmergebot 2025-09-07T07:46:55.2268044Z PR_NUMBER= 2025-09-07T07:46:55.2268173Z DESIRED_CUDA= 2025-09-07T07:46:55.2268320Z GITHUB_RUN_ATTEMPT=1 2025-09-07T07:46:55.2268489Z ANACONDA_PYTHON_VERSION=3.9 2025-09-07T07:46:55.2268700Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-09-07T07:46:55.2268908Z TERM=vt100 2025-09-07T07:46:55.2269049Z INSTALLED_VISION=yes 2025-09-07T07:46:55.2269204Z BRANCH=main 2025-09-07T07:46:55.2269350Z SCCACHE_REGION=us-east-1 2025-09-07T07:46:55.2269526Z OPENSSL_ROOT_DIR=/opt/openssl 2025-09-07T07:46:55.2269710Z CUDA_PATH=/usr/local/cuda 2025-09-07T07:46:55.2270028Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-09-07T07:46:55.2270370Z GITHUB_SERVER_URL=https://github.com 2025-09-07T07:46:55.2270563Z UCC_COMMIT= 2025-09-07T07:46:55.2270789Z REENABLED_ISSUES= 2025-09-07T07:46:55.2270935Z DOCS=yes 2025-09-07T07:46:55.2271071Z SHLVL=1 2025-09-07T07:46:55.2271198Z MAX_JOBS=94 2025-09-07T07:46:55.2271343Z GITHUB_ACTOR_ID=97764156 2025-09-07T07:46:55.2271558Z GITHUB_WORKFLOW_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:46:55.2271794Z GITHUB_REF_NAME=main 2025-09-07T07:46:55.2272028Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-09-07T07:46:55.2272288Z GITHUB_JOB=test 2025-09-07T07:46:55.2272441Z NO_TEST_TIMEOUT=False 2025-09-07T07:46:55.2272600Z TD_DISTRIBUTED=False 2025-09-07T07:46:55.2272767Z GITHUB_REPOSITORY=pytorch/pytorch 2025-09-07T07:46:55.2272956Z GITHUB_RETENTION_DAYS=90 2025-09-07T07:46:55.2273126Z OPENSSL_DIR=/opt/openssl 2025-09-07T07:46:55.2273289Z GITHUB_ACTION_REPOSITORY= 2025-09-07T07:46:55.2273764Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.9/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-09-07T07:46:55.2274260Z GITHUB_BASE_REF= 2025-09-07T07:46:55.2274420Z INSTALLED_ACL= 2025-09-07T07:46:55.2274717Z ARTIFACTS_FILE_SUFFIX=test-inductor_torchbench_perf_cpu_x86_zen-2-4-linux.24xlarge.amd_49775530527 2025-09-07T07:46:55.2275030Z CI=true 2025-09-07T07:46:55.2275179Z GITHUB_REPOSITORY_OWNER=pytorch 2025-09-07T07:46:55.2275405Z RUST_LOG=sccache::server=error 2025-09-07T07:46:55.2275582Z JOB_ID=49775530527 2025-09-07T07:46:55.2275725Z GITHUB_HEAD_REF= 2025-09-07T07:46:55.2275872Z GITHUB_ACTION_REF= 2025-09-07T07:46:55.2276060Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-09-07T07:46:55.2276279Z TEST_SHOWLOCALS=False 2025-09-07T07:46:55.2276467Z GITHUB_WORKFLOW=inductor-perf-nightly-x86-zen 2025-09-07T07:46:55.2276691Z DEBIAN_FRONTEND=noninteractive 2025-09-07T07:46:55.2277070Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_82e2c6f3-282a-484f-91e3-31f2b15ab45b 2025-09-07T07:46:55.2277443Z NO_TD=False 2025-09-07T07:46:55.2277600Z SKIP_SCCACHE_INITIALIZATION=1 2025-09-07T07:46:55.2277794Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-09-07T07:46:55.2277988Z _=/usr/bin/env 2025-09-07T07:46:55.2278207Z ++ python -c 'import site; print(site.getsitepackages()[0])' 2025-09-07T07:46:55.2515059Z + TORCH_INSTALL_DIR=/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch 2025-09-07T07:46:55.2515477Z + TORCH_BIN_DIR=/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/bin 2025-09-07T07:46:55.2515823Z + TORCH_LIB_DIR=/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/lib 2025-09-07T07:46:55.2516165Z + TORCH_TEST_DIR=/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/test 2025-09-07T07:46:55.2516429Z + BUILD_DIR=build 2025-09-07T07:46:55.2516590Z + BUILD_RENAMED_DIR=build_renamed 2025-09-07T07:46:55.2516787Z + BUILD_BIN_DIR=build/bin 2025-09-07T07:46:55.2516952Z + SHARD_NUMBER=2 2025-09-07T07:46:55.2517099Z + NUM_TEST_SHARDS=4 2025-09-07T07:46:55.2517262Z + export TORCH_SERIALIZATION_DEBUG=1 2025-09-07T07:46:55.2517674Z + TORCH_SERIALIZATION_DEBUG=1 2025-09-07T07:46:55.2517863Z + export VALGRIND=ON 2025-09-07T07:46:55.2518015Z + VALGRIND=ON 2025-09-07T07:46:55.2518194Z + [[ linux-jammy-py3.9-gcc11-build == *clang9* ]] 2025-09-07T07:46:55.2518435Z + [[ linux-jammy-py3.9-gcc11-build == *xpu* ]] 2025-09-07T07:46:55.2518640Z + detect_cuda_arch 2025-09-07T07:46:55.2518815Z + [[ linux-jammy-py3.9-gcc11-build == *cuda* ]] 2025-09-07T07:46:55.2519037Z + [[ linux-jammy-py3.9-gcc11-build == *s390x* ]] 2025-09-07T07:46:55.2519235Z + [[ 0 == \1 ]] 2025-09-07T07:46:55.2519375Z + [[ True == \1 ]] 2025-09-07T07:46:55.2519542Z + [[ linux-jammy-py3.9-gcc11-build != *bazel* ]] 2025-09-07T07:46:55.2520227Z ++ realpath build/custom_test_artifacts 2025-09-07T07:46:55.2532281Z + CUSTOM_TEST_ARTIFACT_BUILD_DIR=/var/lib/jenkins/workspace/build/custom_test_artifacts 2025-09-07T07:46:55.2533003Z + [[ -n '' ]] 2025-09-07T07:46:55.2533167Z + echo 'Environment variables' 2025-09-07T07:46:55.2533348Z Environment variables 2025-09-07T07:46:55.2533508Z + env 2025-09-07T07:46:55.2553084Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-09-07T07:46:55.2553391Z CONTINUE_THROUGH_ERROR=True 2025-09-07T07:46:55.2553608Z BUILD_ENVIRONMENT=linux-jammy-py3.9-gcc11-build 2025-09-07T07:46:55.2553990Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-09-07T07:46:55.2554176Z HOSTNAME=de0caba1838c 2025-09-07T07:46:55.2554531Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_82e2c6f3-282a-484f-91e3-31f2b15ab45b 2025-09-07T07:46:55.2554898Z GITHUB_ACTION=__run_2 2025-09-07T07:46:55.2555072Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-09-07T07:46:55.2555262Z GITHUB_RUN_NUMBER=91 2025-09-07T07:46:55.2555444Z TEST_CONFIG=inductor_torchbench_perf_cpu_x86_zen 2025-09-07T07:46:55.2555669Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-09-07T07:46:55.2555877Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-09-07T07:46:55.2556072Z SCCACHE_IDLE_TIMEOUT=0 2025-09-07T07:46:55.2556323Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-09-07T07:46:55.2556520Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-09-07T07:46:55.2556718Z GITHUB_REF_TYPE=branch 2025-09-07T07:46:55.2556902Z BASE_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:46:55.2557099Z XLA_CUDA= 2025-09-07T07:46:55.2557248Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-09-07T07:46:55.2557574Z HUGGING_FACE_HUB_TOKEN=*** 2025-09-07T07:46:55.2557803Z *** 2025-09-07T07:46:55.2557952Z GITHUB_REPOSITORY_ID=65600975 2025-09-07T07:46:55.2558128Z GITHUB_ACTIONS=true 2025-09-07T07:46:55.2558319Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-09-07T07:46:55.2558567Z SHA1=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:46:55.2558795Z GITHUB_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:46:55.2559195Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/inductor-perf-test-nightly-x86-zen.yml@refs/heads/main 2025-09-07T07:46:55.2559554Z UCC_HOME=/usr 2025-09-07T07:46:55.2559706Z TORCH_SERIALIZATION_DEBUG=1 2025-09-07T07:46:55.2559880Z VERBOSE_TEST_LOGS=False 2025-09-07T07:46:55.2560042Z GITHUB_REF=refs/heads/main 2025-09-07T07:46:55.2560206Z SHARD_NUMBER=2 2025-09-07T07:46:55.2560357Z GITHUB_REF_PROTECTED=true 2025-09-07T07:46:55.2560520Z HOME=/var/lib/jenkins 2025-09-07T07:46:55.2560694Z GITHUB_API_URL=https://api.github.com 2025-09-07T07:46:55.2560914Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-09-07T07:46:55.2561100Z UCX_COMMIT= 2025-09-07T07:46:55.2561240Z USE_SYSTEM_NCCL=1 2025-09-07T07:46:55.2561382Z NUM_TEST_SHARDS=4 2025-09-07T07:46:55.2561526Z UCX_HOME=/usr 2025-09-07T07:46:55.2561883Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_82e2c6f3-282a-484f-91e3-31f2b15ab45b 2025-09-07T07:46:55.2562407Z JOB_NAME=inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd) 2025-09-07T07:46:55.2562903Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_82e2c6f3-282a-484f-91e3-31f2b15ab45b 2025-09-07T07:46:55.2563367Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-09-07T07:46:55.2563760Z GITHUB_EVENT_NAME=schedule 2025-09-07T07:46:55.2564155Z DASHBOARD_TAG=training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true 2025-09-07T07:46:55.2564560Z GITHUB_RUN_ID=17525294857 2025-09-07T07:46:55.2564770Z INSTALLED_OPENBLAS= 2025-09-07T07:46:55.2565127Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_82e2c6f3-282a-484f-91e3-31f2b15ab45b 2025-09-07T07:46:55.2565522Z GITHUB_ACTOR=pytorchmergebot 2025-09-07T07:46:55.2565694Z PR_NUMBER= 2025-09-07T07:46:55.2565834Z DESIRED_CUDA= 2025-09-07T07:46:55.2565970Z GITHUB_RUN_ATTEMPT=1 2025-09-07T07:46:55.2566121Z VALGRIND=ON 2025-09-07T07:46:55.2566269Z ANACONDA_PYTHON_VERSION=3.9 2025-09-07T07:46:55.2566478Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-09-07T07:46:55.2566684Z TERM=vt100 2025-09-07T07:46:55.2566828Z INSTALLED_VISION=yes 2025-09-07T07:46:55.2566981Z BRANCH=main 2025-09-07T07:46:55.2567121Z SCCACHE_REGION=us-east-1 2025-09-07T07:46:55.2567379Z OPENSSL_ROOT_DIR=/opt/openssl 2025-09-07T07:46:55.2567561Z CUDA_PATH=/usr/local/cuda 2025-09-07T07:46:55.2567879Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-09-07T07:46:55.2568223Z GITHUB_SERVER_URL=https://github.com 2025-09-07T07:46:55.2568405Z UCC_COMMIT= 2025-09-07T07:46:55.2568545Z REENABLED_ISSUES= 2025-09-07T07:46:55.2568690Z DOCS=yes 2025-09-07T07:46:55.2568820Z SHLVL=1 2025-09-07T07:46:55.2568943Z MAX_JOBS=94 2025-09-07T07:46:55.2569080Z GITHUB_ACTOR_ID=97764156 2025-09-07T07:46:55.2569290Z GITHUB_WORKFLOW_SHA=93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T07:46:55.2569520Z GITHUB_REF_NAME=main 2025-09-07T07:46:55.2569752Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-09-07T07:46:55.2570011Z GITHUB_JOB=test 2025-09-07T07:46:55.2570163Z NO_TEST_TIMEOUT=False 2025-09-07T07:46:55.2570322Z TD_DISTRIBUTED=False 2025-09-07T07:46:55.2570484Z GITHUB_REPOSITORY=pytorch/pytorch 2025-09-07T07:46:55.2570680Z GITHUB_RETENTION_DAYS=90 2025-09-07T07:46:55.2570845Z OPENSSL_DIR=/opt/openssl 2025-09-07T07:46:55.2571000Z GITHUB_ACTION_REPOSITORY= 2025-09-07T07:46:55.2571457Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.9/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-09-07T07:46:55.2571915Z GITHUB_BASE_REF= 2025-09-07T07:46:55.2572063Z INSTALLED_ACL= 2025-09-07T07:46:55.2572359Z ARTIFACTS_FILE_SUFFIX=test-inductor_torchbench_perf_cpu_x86_zen-2-4-linux.24xlarge.amd_49775530527 2025-09-07T07:46:55.2572670Z CI=true 2025-09-07T07:46:55.2572818Z GITHUB_REPOSITORY_OWNER=pytorch 2025-09-07T07:46:55.2573035Z RUST_LOG=sccache::server=error 2025-09-07T07:46:55.2573209Z JOB_ID=49775530527 2025-09-07T07:46:55.2573352Z GITHUB_HEAD_REF= 2025-09-07T07:46:55.2573497Z GITHUB_ACTION_REF= 2025-09-07T07:46:55.2573682Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-09-07T07:46:55.2573898Z TEST_SHOWLOCALS=False 2025-09-07T07:46:55.2574092Z GITHUB_WORKFLOW=inductor-perf-nightly-x86-zen 2025-09-07T07:46:55.2574309Z DEBIAN_FRONTEND=noninteractive 2025-09-07T07:46:55.2574680Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_82e2c6f3-282a-484f-91e3-31f2b15ab45b 2025-09-07T07:46:55.2575050Z NO_TD=False 2025-09-07T07:46:55.2575195Z SKIP_SCCACHE_INITIALIZATION=1 2025-09-07T07:46:55.2575387Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-09-07T07:46:55.2575581Z _=/usr/bin/env 2025-09-07T07:46:55.2575728Z + echo 'Testing pytorch' 2025-09-07T07:46:55.2575888Z Testing pytorch 2025-09-07T07:46:55.2576055Z + export LANG=C.UTF-8 2025-09-07T07:46:55.2576210Z + LANG=C.UTF-8 2025-09-07T07:46:55.2576345Z + PR_NUMBER= 2025-09-07T07:46:55.2576551Z + [[ inductor_torchbench_perf_cpu_x86_zen == \d\e\f\a\u\l\t ]] 2025-09-07T07:46:55.2576848Z + [[ inductor_torchbench_perf_cpu_x86_zen == \d\i\s\t\r\i\b\u\t\e\d ]] 2025-09-07T07:46:55.2577131Z + [[ inductor_torchbench_perf_cpu_x86_zen == \s\l\o\w ]] 2025-09-07T07:46:55.2577522Z + [[ linux-jammy-py3.9-gcc11-build == *slow-gradcheck* ]] 2025-09-07T07:46:55.2577813Z + [[ linux-jammy-py3.9-gcc11-build == *cuda* ]] 2025-09-07T07:46:55.2578031Z + [[ linux-jammy-py3.9-gcc11-build == *rocm* ]] 2025-09-07T07:46:55.2578249Z + [[ linux-jammy-py3.9-gcc11-build == *xpu* ]] 2025-09-07T07:46:55.2578490Z + [[ inductor_torchbench_perf_cpu_x86_zen == *crossref* ]] 2025-09-07T07:46:55.2578724Z + [[ linux-jammy-py3.9-gcc11-build == *rocm* ]] 2025-09-07T07:46:55.2578936Z + [[ linux-jammy-py3.9-gcc11-build == *xpu* ]] 2025-09-07T07:46:55.2579161Z + [[ linux-jammy-py3.9-gcc11-build != *-bazel-* ]] 2025-09-07T07:46:55.2579380Z + pip_install ninja==1.10.2 2025-09-07T07:46:55.2579608Z + pip_install_pkg='python3 -m pip install --progress-bar off' 2025-09-07T07:46:55.2579890Z + python3 -m pip install --progress-bar off ninja==1.10.2 2025-09-07T07:46:55.6392898Z Collecting ninja==1.10.2 2025-09-07T07:46:55.6469398Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl.metadata (5.0 kB) 2025-09-07T07:46:55.6564004Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl (108 kB) 2025-09-07T07:46:56.3377600Z Installing collected packages: ninja 2025-09-07T07:46:56.3377920Z Attempting uninstall: ninja 2025-09-07T07:46:56.3385316Z Found existing installation: ninja 1.11.1.3 2025-09-07T07:46:56.3406848Z Uninstalling ninja-1.11.1.3: 2025-09-07T07:46:56.3478356Z Successfully uninstalled ninja-1.11.1.3 2025-09-07T07:46:56.4054755Z Successfully installed ninja-1.10.2 2025-09-07T07:46:56.4803903Z + export PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.9/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-09-07T07:46:56.4804840Z + PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.9/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-09-07T07:46:56.4805470Z + [[ linux-jammy-py3.9-gcc11-build == *aarch64* ]] 2025-09-07T07:46:56.4805732Z + [[ linux-jammy-py3.9-gcc11-build == *asan* ]] 2025-09-07T07:46:56.4805962Z + [[ linux-jammy-py3.9-gcc11-build == *-debug* ]] 2025-09-07T07:46:56.4806201Z + [[ linux-jammy-py3.9-gcc11-build != *-bazel-* ]] 2025-09-07T07:46:56.4806538Z + echo 'We are not in debug mode: linux-jammy-py3.9-gcc11-build. Expect the assertion to pass' 2025-09-07T07:46:56.4806943Z We are not in debug mode: linux-jammy-py3.9-gcc11-build. Expect the assertion to pass 2025-09-07T07:46:56.4809522Z + cd test 2025-09-07T07:46:56.4809754Z + python -c 'import torch; torch._C._crash_if_debug_asserts_fail(424242)' 2025-09-07T07:46:56.7616638Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:46:56.7617590Z import pynvml # type: ignore[import] 2025-09-07T07:46:57.5846049Z + [[ inductor_torchbench_perf_cpu_x86_zen == \n\o\g\p\u\_\N\O\_\A\V\X\2 ]] 2025-09-07T07:46:57.5846446Z + [[ inductor_torchbench_perf_cpu_x86_zen == \n\o\g\p\u\_\A\V\X\5\1\2 ]] 2025-09-07T07:46:57.5846815Z + [[ inductor_torchbench_perf_cpu_x86_zen == \l\e\g\a\c\y\_\n\v\i\d\i\a\_\d\r\i\v\e\r ]] 2025-09-07T07:46:57.5847965Z + DYNAMO_BENCHMARK_FLAGS=() 2025-09-07T07:46:57.5849668Z + [[ inductor_torchbench_perf_cpu_x86_zen == *pr_time_benchmarks* ]] 2025-09-07T07:46:57.5849996Z + [[ inductor_torchbench_perf_cpu_x86_zen == *dynamo_eager* ]] 2025-09-07T07:46:57.5850272Z + [[ inductor_torchbench_perf_cpu_x86_zen == *aot_eager* ]] 2025-09-07T07:46:57.5850538Z + [[ inductor_torchbench_perf_cpu_x86_zen == *aot_inductor* ]] 2025-09-07T07:46:57.5850832Z + [[ inductor_torchbench_perf_cpu_x86_zen == *max_autotune_inductor* ]] 2025-09-07T07:46:57.5851130Z + [[ inductor_torchbench_perf_cpu_x86_zen == *inductor* ]] 2025-09-07T07:46:57.5851388Z + [[ inductor_torchbench_perf_cpu_x86_zen != *perf* ]] 2025-09-07T07:46:57.5852100Z + [[ inductor_torchbench_perf_cpu_x86_zen == *dynamic* ]] 2025-09-07T07:46:57.5852339Z + [[ inductor_torchbench_perf_cpu_x86_zen == *cpu* ]] 2025-09-07T07:46:57.5852579Z + DYNAMO_BENCHMARK_FLAGS+=(--device cpu) 2025-09-07T07:46:57.6083233Z + [[ linux-jammy-py3.9-gcc11-build == *libtorch* ]] 2025-09-07T07:46:57.6083501Z + [[ linux-jammy-py3.9-gcc11-build == *-bazel-* ]] 2025-09-07T07:46:57.6086559Z + cd test 2025-09-07T07:46:57.6087338Z + python -c 'import torch; print(torch.__config__.show())' 2025-09-07T07:46:57.8845389Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:46:57.8846225Z import pynvml # type: ignore[import] 2025-09-07T07:46:58.5109505Z PyTorch built with: 2025-09-07T07:46:58.5109781Z - GCC 11.4 2025-09-07T07:46:58.5110421Z - C++ Version: 201703 2025-09-07T07:46:58.5110788Z - Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-09-07T07:46:58.5111230Z - Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-09-07T07:46:58.5111508Z - OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-09-07T07:46:58.5111730Z - LAPACK is enabled (usually provided by MKL) 2025-09-07T07:46:58.5111929Z - NNPACK is enabled 2025-09-07T07:46:58.5112102Z - CPU capability usage: AVX512 2025-09-07T07:46:58.5114869Z - Build settings: BLAS_INFO=mkl, BUILD_TYPE=Release, COMMIT_SHA=93fb23d6fae7c4e82c4239a1033e522088742634, CXX_COMPILER=/opt/cache/bin/c++, CXX_FLAGS= -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOCUPTI -DLIBKINETO_NOROCTRACER -DLIBKINETO_NOXPUPTI=ON -DUSE_FBGEMM -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -DC10_NODEPRECATED -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=range-loop-construct -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-unknown-pragmas -Wno-unused-parameter -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wsuggest-override -Wno-psabi -Wno-error=old-style-cast -faligned-new -Werror -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow, LAPACK_INFO=mkl, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, TORCH_VERSION=2.9.0, USE_CUDA=OFF, USE_CUDNN=OFF, USE_CUSPARSELT=OFF, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_GLOO=ON, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=OFF, USE_NCCL=OFF, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, USE_ROCM_KERNEL_ASSERT=OFF, USE_XCCL=OFF, USE_XPU=OFF, 2025-09-07T07:46:58.5117616Z 2025-09-07T07:46:58.6954765Z + cd test 2025-09-07T07:46:58.6955037Z + python -c 'import torch; print(torch.__config__.parallel_info())' 2025-09-07T07:46:58.9679541Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:46:58.9680381Z import pynvml # type: ignore[import] 2025-09-07T07:46:59.5971477Z ATen/Parallel: 2025-09-07T07:46:59.5971754Z at::get_num_threads() : 96 2025-09-07T07:46:59.5971958Z at::get_num_interop_threads() : 96 2025-09-07T07:46:59.5972168Z OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-09-07T07:46:59.5972362Z omp_get_max_threads() : 96 2025-09-07T07:46:59.5972711Z Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-09-07T07:46:59.5973064Z mkl_get_max_threads() : 96 2025-09-07T07:46:59.5973315Z Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-09-07T07:46:59.5973593Z std::thread::hardware_concurrency() : 96 2025-09-07T07:46:59.5973797Z Environment variables: 2025-09-07T07:46:59.5973975Z OMP_NUM_THREADS : [not set] 2025-09-07T07:46:59.5974609Z MKL_NUM_THREADS : [not set] 2025-09-07T07:46:59.5974789Z ATen parallel backend: OpenMP 2025-09-07T07:46:59.5974906Z 2025-09-07T07:46:59.7787712Z + [[ inductor_torchbench_perf_cpu_x86_zen == *numpy_2* ]] 2025-09-07T07:46:59.7788051Z + [[ linux-jammy-py3.9-gcc11-build == *aarch64* ]] 2025-09-07T07:46:59.7788326Z + [[ inductor_torchbench_perf_cpu_x86_zen == *backward* ]] 2025-09-07T07:46:59.7788593Z + [[ inductor_torchbench_perf_cpu_x86_zen == *xla* ]] 2025-09-07T07:46:59.7788832Z + [[ inductor_torchbench_perf_cpu_x86_zen == *vllm* ]] 2025-09-07T07:46:59.7789090Z + [[ inductor_torchbench_perf_cpu_x86_zen == *executorch* ]] 2025-09-07T07:46:59.7789393Z + [[ inductor_torchbench_perf_cpu_x86_zen == \j\i\t\_\l\e\g\a\c\y ]] 2025-09-07T07:46:59.7789676Z + [[ linux-jammy-py3.9-gcc11-build == *libtorch* ]] 2025-09-07T07:46:59.7789928Z + [[ inductor_torchbench_perf_cpu_x86_zen == distributed ]] 2025-09-07T07:46:59.7790219Z + [[ inductor_torchbench_perf_cpu_x86_zen == *operator_benchmark* ]] 2025-09-07T07:46:59.7790939Z + [[ inductor_torchbench_perf_cpu_x86_zen == *inductor_distributed* ]] 2025-09-07T07:46:59.7791253Z + [[ inductor_torchbench_perf_cpu_x86_zen == *inductor-halide* ]] 2025-09-07T07:46:59.7791560Z + [[ inductor_torchbench_perf_cpu_x86_zen == *inductor-triton-cpu* ]] 2025-09-07T07:46:59.7791881Z + [[ inductor_torchbench_perf_cpu_x86_zen == *inductor-micro-benchmark* ]] 2025-09-07T07:46:59.7792189Z + [[ inductor_torchbench_perf_cpu_x86_zen == *huggingface* ]] 2025-09-07T07:46:59.7792453Z + [[ inductor_torchbench_perf_cpu_x86_zen == *timm* ]] 2025-09-07T07:46:59.7792707Z + [[ inductor_torchbench_perf_cpu_x86_zen == cachebench ]] 2025-09-07T07:46:59.7792967Z + [[ inductor_torchbench_perf_cpu_x86_zen == verify_cachebench ]] 2025-09-07T07:46:59.7793245Z + [[ inductor_torchbench_perf_cpu_x86_zen == *torchbench* ]] 2025-09-07T07:46:59.7793469Z + install_torchaudio 2025-09-07T07:46:59.7793634Z + local commit 2025-09-07T07:46:59.7793791Z ++ get_pinned_commit audio 2025-09-07T07:46:59.7793985Z ++ cat .github/ci_commit_pins/audio.txt 2025-09-07T07:46:59.8132696Z + commit=2e300559e4e123928a22187b8f59a5b56f57ddc8 2025-09-07T07:46:59.8133144Z + pip_build_and_install git+https://github.com/pytorch/audio.git@2e300559e4e123928a22187b8f59a5b56f57ddc8 dist/audio 2025-09-07T07:46:59.8134129Z + local build_target=git+https://github.com/pytorch/audio.git@2e300559e4e123928a22187b8f59a5b56f57ddc8 2025-09-07T07:46:59.8134457Z + local wheel_dir=dist/audio 2025-09-07T07:46:59.8134634Z + local found_whl=0 2025-09-07T07:46:59.8134802Z + for file in "${wheel_dir}"/*.whl 2025-09-07T07:46:59.8135000Z + [[ -f dist/audio/*.whl ]] 2025-09-07T07:46:59.8135165Z + '[' 0 == 0 ']' 2025-09-07T07:46:59.8135613Z + python3 -m pip wheel --no-build-isolation --no-deps --no-use-pep517 -w dist/audio git+https://github.com/pytorch/audio.git@2e300559e4e123928a22187b8f59a5b56f57ddc8 2025-09-07T07:47:00.1037714Z Collecting git+https://github.com/pytorch/audio.git@2e300559e4e123928a22187b8f59a5b56f57ddc8 2025-09-07T07:47:00.1040994Z Cloning https://github.com/pytorch/audio.git (to revision 2e300559e4e123928a22187b8f59a5b56f57ddc8) to /tmp/pip-req-build-6hlkg3l_ 2025-09-07T07:47:00.1928971Z Running command git clone --filter=blob:none --quiet https://github.com/pytorch/audio.git /tmp/pip-req-build-6hlkg3l_ 2025-09-07T07:47:02.7461600Z Running command git rev-parse -q --verify 'sha^2e300559e4e123928a22187b8f59a5b56f57ddc8' 2025-09-07T07:47:02.7505540Z Running command git fetch -q https://github.com/pytorch/audio.git 2e300559e4e123928a22187b8f59a5b56f57ddc8 2025-09-07T07:47:02.8609228Z Running command git checkout -q 2e300559e4e123928a22187b8f59a5b56f57ddc8 2025-09-07T07:47:03.2290538Z Resolved https://github.com/pytorch/audio.git to commit 2e300559e4e123928a22187b8f59a5b56f57ddc8 2025-09-07T07:47:03.2290984Z Running command git submodule update --init --recursive -q 2025-09-07T07:47:04.6178398Z Preparing metadata (setup.py) ... [?25l- \ done 2025-09-07T07:47:04.6216702Z [?25hBuilding wheels for collected packages: torchaudio 2025-09-07T07:47:04.6323532Z  DEPRECATION: Building 'torchaudio' using the legacy setup.py bdist_wheel mechanism, which will be removed in a future version. pip 25.3 will enforce this behaviour change. A possible replacement is to use the standardized build interface by setting the `--use-pep517` option, (possibly combined with `--no-build-isolation`), or adding a `pyproject.toml` file to the source tree of 'torchaudio'. Discussion can be found at https://github.com/pypa/pip/issues/6334 2025-09-07T07:47:26.4696591Z  Building wheel for torchaudio (setup.py) ... [?25l- \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | done 2025-09-07T07:47:26.4710682Z [?25h Created wheel for torchaudio: filename=torchaudio-2.8.0a0+2e30055-cp39-cp39-linux_x86_64.whl size=491187 sha256=546ff9170a0d92d409fcd0637612dab05fc5407ccdb6fcf76c21b3ed9755e7dc 2025-09-07T07:47:26.4712166Z Stored in directory: /var/lib/jenkins/.cache/pip/wheels/18/53/66/85c241150a0c0641633cdbd6ae534ee172017679f2ef448df1 2025-09-07T07:47:26.4742266Z Successfully built torchaudio 2025-09-07T07:47:26.5920733Z + for file in "${wheel_dir}"/*.whl 2025-09-07T07:47:26.5921095Z + pip_install_whl dist/audio/torchaudio-2.8.0a0+2e30055-cp39-cp39-linux_x86_64.whl 2025-09-07T07:47:26.5921513Z + args=('dist/audio/torchaudio-2.8.0a0+2e30055-cp39-cp39-linux_x86_64.whl') 2025-09-07T07:47:26.5921780Z + local args 2025-09-07T07:47:26.5922025Z + [[ dist/audio/torchaudio-2.8.0a0+2e30055-cp39-cp39-linux_x86_64.whl == *\ * ]] 2025-09-07T07:47:26.5922328Z + for path in "${args[@]}" 2025-09-07T07:47:26.5922611Z + echo 'Installing dist/audio/torchaudio-2.8.0a0+2e30055-cp39-cp39-linux_x86_64.whl' 2025-09-07T07:47:26.5923006Z Installing dist/audio/torchaudio-2.8.0a0+2e30055-cp39-cp39-linux_x86_64.whl 2025-09-07T07:47:26.5923447Z + python3 -mpip install --no-index --no-deps dist/audio/torchaudio-2.8.0a0+2e30055-cp39-cp39-linux_x86_64.whl 2025-09-07T07:47:26.8920096Z Processing ./dist/audio/torchaudio-2.8.0a0+2e30055-cp39-cp39-linux_x86_64.whl 2025-09-07T07:47:26.8969110Z Installing collected packages: torchaudio 2025-09-07T07:47:27.0569075Z Successfully installed torchaudio-2.8.0a0+2e30055 2025-09-07T07:47:27.0877394Z + install_torchvision 2025-09-07T07:47:27.0877591Z + local orig_preload 2025-09-07T07:47:27.0877750Z + local commit 2025-09-07T07:47:27.0882143Z ++ get_pinned_commit vision 2025-09-07T07:47:27.0882341Z ++ cat .github/ci_commit_pins/vision.txt 2025-09-07T07:47:27.0899156Z + commit=966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-09-07T07:47:27.0899373Z + orig_preload= 2025-09-07T07:47:27.0899529Z + '[' -n '' ']' 2025-09-07T07:47:27.0899712Z + [[ linux-jammy-py3.9-gcc11-build == *cuda* ]] 2025-09-07T07:47:27.0900154Z + pip_build_and_install git+https://github.com/pytorch/vision.git@966da7e46f65d6d49df3e31214470a4fe5cc8e66 dist/vision 2025-09-07T07:47:27.0900682Z + local build_target=git+https://github.com/pytorch/vision.git@966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-09-07T07:47:27.0901083Z + local wheel_dir=dist/vision 2025-09-07T07:47:27.0901276Z + local found_whl=0 2025-09-07T07:47:27.0901440Z + for file in "${wheel_dir}"/*.whl 2025-09-07T07:47:27.0901631Z + [[ -f dist/vision/*.whl ]] 2025-09-07T07:47:27.0901807Z + '[' 0 == 0 ']' 2025-09-07T07:47:27.0902256Z + python3 -m pip wheel --no-build-isolation --no-deps --no-use-pep517 -w dist/vision git+https://github.com/pytorch/vision.git@966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-09-07T07:47:27.3830180Z Collecting git+https://github.com/pytorch/vision.git@966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-09-07T07:47:27.3833979Z Cloning https://github.com/pytorch/vision.git (to revision 966da7e46f65d6d49df3e31214470a4fe5cc8e66) to /tmp/pip-req-build-75k5m33q 2025-09-07T07:47:27.3885071Z Running command git clone --filter=blob:none --quiet https://github.com/pytorch/vision.git /tmp/pip-req-build-75k5m33q 2025-09-07T07:47:28.7561794Z Running command git rev-parse -q --verify 'sha^966da7e46f65d6d49df3e31214470a4fe5cc8e66' 2025-09-07T07:47:28.7606517Z Running command git fetch -q https://github.com/pytorch/vision.git 966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-09-07T07:47:28.9147958Z Running command git checkout -q 966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-09-07T07:47:29.2015697Z Resolved https://github.com/pytorch/vision.git to commit 966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-09-07T07:47:30.6911342Z Preparing metadata (setup.py) ... [?25l- \ | / done 2025-09-07T07:47:30.6951240Z [?25hBuilding wheels for collected packages: torchvision 2025-09-07T07:47:30.7059994Z  DEPRECATION: Building 'torchvision' using the legacy setup.py bdist_wheel mechanism, which will be removed in a future version. pip 25.3 will enforce this behaviour change. A possible replacement is to use the standardized build interface by setting the `--use-pep517` option, (possibly combined with `--no-build-isolation`), or adding a `pyproject.toml` file to the source tree of 'torchvision'. Discussion can be found at https://github.com/pypa/pip/issues/6334 2025-09-07T07:47:59.2299825Z  Building wheel for torchvision (setup.py) ... [?25l- \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | done 2025-09-07T07:47:59.2322588Z [?25h Created wheel for torchvision: filename=torchvision-0.22.0a0+966da7e-cp39-cp39-linux_x86_64.whl size=1301751 sha256=9573c154c3503860798ceb40afb0069782afb98584a9dfd18ea8c8720804076c 2025-09-07T07:47:59.2323377Z Stored in directory: /var/lib/jenkins/.cache/pip/wheels/33/6d/2f/9f3e65c401a351a98a00d9d72c4434fdbd3e10256b2d832157 2025-09-07T07:47:59.2361977Z Successfully built torchvision 2025-09-07T07:47:59.3252004Z + for file in "${wheel_dir}"/*.whl 2025-09-07T07:47:59.3252471Z + pip_install_whl dist/vision/torchvision-0.22.0a0+966da7e-cp39-cp39-linux_x86_64.whl 2025-09-07T07:47:59.3264545Z + args=('dist/vision/torchvision-0.22.0a0+966da7e-cp39-cp39-linux_x86_64.whl') 2025-09-07T07:47:59.3264855Z + local args 2025-09-07T07:47:59.3265125Z + [[ dist/vision/torchvision-0.22.0a0+966da7e-cp39-cp39-linux_x86_64.whl == *\ * ]] 2025-09-07T07:47:59.3265434Z + for path in "${args[@]}" 2025-09-07T07:47:59.3265722Z + echo 'Installing dist/vision/torchvision-0.22.0a0+966da7e-cp39-cp39-linux_x86_64.whl' 2025-09-07T07:47:59.3266114Z Installing dist/vision/torchvision-0.22.0a0+966da7e-cp39-cp39-linux_x86_64.whl 2025-09-07T07:47:59.3266566Z + python3 -mpip install --no-index --no-deps dist/vision/torchvision-0.22.0a0+966da7e-cp39-cp39-linux_x86_64.whl 2025-09-07T07:47:59.6258729Z Processing ./dist/vision/torchvision-0.22.0a0+966da7e-cp39-cp39-linux_x86_64.whl 2025-09-07T07:47:59.6329770Z Installing collected packages: torchvision 2025-09-07T07:48:00.0013950Z Successfully installed torchvision-0.22.0a0+966da7e 2025-09-07T07:48:00.0353989Z + '[' -n '' ']' 2025-09-07T07:48:00.0354175Z + id=1 2025-09-07T07:48:00.0354350Z + pip_install opencv-python==4.8.0.74 2025-09-07T07:48:00.0354629Z + pip_install_pkg='python3 -m pip install --progress-bar off' 2025-09-07T07:48:00.0354970Z + python3 -m pip install --progress-bar off opencv-python==4.8.0.74 2025-09-07T07:48:00.3989887Z Collecting opencv-python==4.8.0.74 2025-09-07T07:48:00.4230357Z Downloading opencv_python-4.8.0.74-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (19 kB) 2025-09-07T07:48:00.4287234Z Requirement already satisfied: numpy>=1.17.0 in /opt/conda/envs/py_3.9/lib/python3.9/site-packages (from opencv-python==4.8.0.74) (1.22.4) 2025-09-07T07:48:00.4395421Z Downloading opencv_python-4.8.0.74-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (61.7 MB) 2025-09-07T07:48:01.4921811Z Installing collected packages: opencv-python 2025-09-07T07:48:01.4922084Z Attempting uninstall: opencv-python 2025-09-07T07:48:01.4932792Z Found existing installation: opencv-python 4.11.0.86 2025-09-07T07:48:01.5010694Z Uninstalling opencv-python-4.11.0.86: 2025-09-07T07:48:01.6203704Z Successfully uninstalled opencv-python-4.11.0.86 2025-09-07T07:48:02.3538752Z Successfully installed opencv-python-4.8.0.74 2025-09-07T07:48:02.4433084Z + [[ inductor_torchbench_perf_cpu_x86_zen == *inductor_torchbench_smoketest_perf* ]] 2025-09-07T07:48:02.4433528Z + [[ inductor_torchbench_perf_cpu_x86_zen == *inductor_torchbench_cpu_smoketest_perf* ]] 2025-09-07T07:48:02.4433910Z + [[ inductor_torchbench_perf_cpu_x86_zen == *torchbench_gcp_smoketest* ]] 2025-09-07T07:48:02.4434217Z + [[ inductor_torchbench_perf_cpu_x86_zen != *cpu* ]] 2025-09-07T07:48:02.4434449Z + PYTHONPATH=/torchbench 2025-09-07T07:48:02.4434652Z + test_dynamo_benchmark torchbench 1 2025-09-07T07:48:02.4437596Z ++ pwd 2025-09-07T07:48:02.4440491Z + TEST_REPORTS_DIR=/var/lib/jenkins/workspace/test/test-reports 2025-09-07T07:48:02.4440766Z + local suite=torchbench 2025-09-07T07:48:02.4440933Z + shift 2025-09-07T07:48:02.4441066Z + local shard_id=1 2025-09-07T07:48:02.4441221Z + shift 2025-09-07T07:48:02.4441425Z + [[ inductor_torchbench_perf_cpu_x86_zen == *perf_compare* ]] 2025-09-07T07:48:02.4441700Z + [[ inductor_torchbench_perf_cpu_x86_zen == *perf* ]] 2025-09-07T07:48:02.4442389Z + [[ inductor_torchbench_perf_cpu_x86_zen == *b200* ]] 2025-09-07T07:48:02.4443669Z + test_single_dynamo_benchmark dashboard torchbench 1 2025-09-07T07:48:02.4446586Z ++ pwd 2025-09-07T07:48:02.4449574Z + TEST_REPORTS_DIR=/var/lib/jenkins/workspace/test/test-reports 2025-09-07T07:48:02.4449951Z + mkdir -p /var/lib/jenkins/workspace/test/test-reports 2025-09-07T07:48:02.4471368Z + local name=dashboard 2025-09-07T07:48:02.4471576Z + shift 2025-09-07T07:48:02.4471738Z + local suite=torchbench 2025-09-07T07:48:02.4471979Z + shift 2025-09-07T07:48:02.4472119Z + local shard_id=1 2025-09-07T07:48:02.4472271Z + shift 2025-09-07T07:48:02.4472420Z + partition_flags=() 2025-09-07T07:48:02.4472594Z + local partition_flags 2025-09-07T07:48:02.4472767Z + [[ -n 4 ]] 2025-09-07T07:48:02.4472907Z + [[ -n 1 ]] 2025-09-07T07:48:02.4473174Z + partition_flags=(--total-partitions "$NUM_TEST_SHARDS" --partition-id "$shard_id") 2025-09-07T07:48:02.4473542Z + [[ inductor_torchbench_perf_cpu_x86_zen == *perf_compare* ]] 2025-09-07T07:48:02.4473885Z + [[ inductor_torchbench_perf_cpu_x86_zen == *perf* ]] 2025-09-07T07:48:02.4474223Z + test_perf_for_dashboard torchbench --device cpu --total-partitions 4 --partition-id 1 2025-09-07T07:48:02.4476916Z ++ pwd 2025-09-07T07:48:02.4479734Z + TEST_REPORTS_DIR=/var/lib/jenkins/workspace/test/test-reports 2025-09-07T07:48:02.4480021Z + mkdir -p /var/lib/jenkins/workspace/test/test-reports 2025-09-07T07:48:02.4499104Z + local suite=torchbench 2025-09-07T07:48:02.4499267Z + shift 2025-09-07T07:48:02.4499407Z + local backend=inductor 2025-09-07T07:48:02.4499573Z + modes=() 2025-09-07T07:48:02.4499717Z + local modes 2025-09-07T07:48:02.4500127Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *training-true* ]] 2025-09-07T07:48:02.4500819Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *inference-true* ]] 2025-09-07T07:48:02.4501251Z + modes+=(inference) 2025-09-07T07:48:02.4501439Z + targets=('accuracy' 'performance') 2025-09-07T07:48:02.4501632Z + local targets 2025-09-07T07:48:02.4501779Z + local device=cuda 2025-09-07T07:48:02.4501964Z + [[ inductor_torchbench_perf_cpu_x86_zen == *cpu* ]] 2025-09-07T07:48:02.4502225Z + [[ inductor_torchbench_perf_cpu_x86_zen == *cpu_x86_zen* ]] 2025-09-07T07:48:02.4502446Z + device=cpu_x86_zen 2025-09-07T07:48:02.4502614Z + test_inductor_set_cpu_affinity 2025-09-07T07:48:02.4504533Z ++ find /usr/lib -name libjemalloc.so.2 2025-09-07T07:48:02.4860731Z + JEMALLOC_LIB=/usr/lib/x86_64-linux-gnu/libjemalloc.so.2 2025-09-07T07:48:02.4861039Z + export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libjemalloc.so.2: 2025-09-07T07:48:02.4861312Z + LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libjemalloc.so.2: 2025-09-07T07:48:02.4861718Z + export MALLOC_CONF=oversize_threshold:1,background_thread:true,metadata_thp:auto,dirty_decay_ms:-1,muzzy_decay_ms:-1 2025-09-07T07:48:02.4862239Z + MALLOC_CONF=oversize_threshold:1,background_thread:true,metadata_thp:auto,dirty_decay_ms:-1,muzzy_decay_ms:-1 2025-09-07T07:48:02.4863892Z + [[ inductor_torchbench_perf_cpu_x86_zen != *aarch64* ]] 2025-09-07T07:48:02.4869632Z +++ which python 2025-09-07T07:48:02.4894814Z ++ dirname /opt/conda/envs/py_3.9/bin/python 2025-09-07T07:48:02.4925780Z + IOMP_LIB=/opt/conda/envs/py_3.9/bin/../lib/libiomp5.so 2025-09-07T07:48:02.4926248Z + export LD_PRELOAD=/opt/conda/envs/py_3.9/bin/../lib/libiomp5.so:/usr/lib/x86_64-linux-gnu/libjemalloc.so.2: 2025-09-07T07:48:02.4926719Z + LD_PRELOAD=/opt/conda/envs/py_3.9/bin/../lib/libiomp5.so:/usr/lib/x86_64-linux-gnu/libjemalloc.so.2: 2025-09-07T07:48:02.4927144Z + export KMP_AFFINITY=granularity=fine,compact,1,0 2025-09-07T07:48:02.4927407Z + KMP_AFFINITY=granularity=fine,compact,1,0 2025-09-07T07:48:02.4927628Z + export KMP_BLOCKTIME=1 2025-09-07T07:48:02.4927795Z + KMP_BLOCKTIME=1 2025-09-07T07:48:02.4930980Z ++ nproc 2025-09-07T07:48:02.4964817Z + cpus=96 2025-09-07T07:48:02.4971936Z ++ lscpu 2025-09-07T07:48:02.4973633Z ++ grep 'Thread(s) per core:' 2025-09-07T07:48:02.4974851Z ++ awk '{print $4}' 2025-09-07T07:48:02.5581609Z + thread_per_core=1 2025-09-07T07:48:02.5581806Z + cores=96 2025-09-07T07:48:02.5582004Z + [[ inductor_torchbench_perf_cpu_x86_zen == *aarch64* ]] 2025-09-07T07:48:02.5582239Z + export OMP_NUM_THREADS=96 2025-09-07T07:48:02.5582413Z + OMP_NUM_THREADS=96 2025-09-07T07:48:02.5586272Z ++ python -c 'import os; print(min(os.sched_getaffinity(0)))' 2025-09-07T07:48:02.5844414Z + start_cpu=0 2025-09-07T07:48:02.5848816Z ++ python -c 'import os; print(max(os.sched_getaffinity(0)))' 2025-09-07T07:48:02.6112074Z + end_cpu=94 2025-09-07T07:48:02.6112262Z + export 'TASKSET=taskset -c 0-94' 2025-09-07T07:48:02.6112474Z + TASKSET='taskset -c 0-94' 2025-09-07T07:48:02.6112668Z + for mode in "${modes[@]}" 2025-09-07T07:48:02.6112834Z + [[ inference == \i\n\f\e\r\e\n\c\e ]] 2025-09-07T07:48:02.6113038Z + [[ cpu_x86_zen == \c\p\u\_\x\8\6 ]] 2025-09-07T07:48:02.6113229Z + dtype=bfloat16 2025-09-07T07:48:02.6113387Z + for target in "${targets[@]}" 2025-09-07T07:48:02.6113588Z + target_flag=('--accuracy') 2025-09-07T07:48:02.6113766Z + local target_flag 2025-09-07T07:48:02.6113931Z + [[ accuracy == \p\e\r\f\o\r\m\a\n\c\e ]] 2025-09-07T07:48:02.6114128Z + [[ accuracy == \a\c\c\u\r\a\c\y ]] 2025-09-07T07:48:02.6114329Z + target_flag+=(--no-translation-validation) 2025-09-07T07:48:02.6114796Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *freezing-true* ]] 2025-09-07T07:48:02.6115477Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *default-true* ]] 2025-09-07T07:48:02.6116661Z + taskset -c 0-94 python benchmarks/dynamo/torchbench.py --accuracy --no-translation-validation --inference --bfloat16 --backend inductor --disable-cudagraphs --device cpu --total-partitions 4 --partition-id 1 --output /var/lib/jenkins/workspace/test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.csv 2025-09-07T07:48:03.0112174Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:48:03.0113066Z import pynvml # type: ignore[import] 2025-09-07T07:48:05.5005237Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:48:05.5006087Z import pynvml # type: ignore[import] 2025-09-07T07:48:07.5021202Z 2025-09-07T07:48:12.7974006Z loading model: 0it [00:00, ?it/s] 2025-09-07T07:48:12.7974360Z loading model: 0it [00:05, ?it/s] 2025-09-07T07:48:12.8250561Z cpu eval dlrm 2025-09-07T07:48:12.9586907Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:48:12.9867915Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:48:13.0375168Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:48:28.5256210Z pass 2025-09-07T07:48:28.5256607Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T07:48:30.8274783Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T07:48:30.8275660Z import pynvml # type: ignore[import] 2025-09-07T07:48:32.8324119Z 2025-09-07T07:48:34.1277714Z loading model: 0it [00:00, ?it/s]Downloading https://doctr-static.mindee.com/models?id=v0.7.0/db_resnet50-79bd7d70.pt&src=0 to /var/lib/jenkins/.cache/doctr/models/db_resnet50-79bd7d70.pt 2025-09-07T07:48:34.3819083Z 2025-09-07T07:48:34.3819092Z 2025-09-07T07:48:34.4820760Z 0% 0/102021912 [00:00 2025-09-07T08:03:06.8490818Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:03:06.8491011Z 2025-09-07T08:03:06.8491121Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8491582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8492010Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8492425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:03:06.8492916Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:03:06.8493414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:03:06.8493913Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:03:06.8494100Z 2025-09-07T08:03:06.8494194Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8494669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8495123Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8495544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:03:06.8496075Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:03:06.8496592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:03:06.8497097Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:03:06.8497293Z 2025-09-07T08:03:06.8497392Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8497946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8498466Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8498879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:03:06.8499374Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:03:06.8499868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:03:06.8500364Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:03:06.8500548Z 2025-09-07T08:03:06.8500647Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8501091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8501524Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8501996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:03:06.8502537Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8503207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:03:06.8503711Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8503900Z 2025-09-07T08:03:06.8504003Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8504477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8504918Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8505351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:03:06.8505830Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8506318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:03:06.8506803Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8506979Z 2025-09-07T08:03:06.8507074Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8507533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8507959Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8508377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:03:06.8508857Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8509334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:03:06.8509818Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8509987Z 2025-09-07T08:03:06.8510084Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8510532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8510961Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8511387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:03:06.8511868Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8512347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:03:06.8512958Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8513135Z 2025-09-07T08:03:06.8513228Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8513684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:03:06.8514118Z logits = self.prob_head(feat_concat) 2025-09-07T08:03:06.8514254Z 2025-09-07T08:03:06.8514342Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8514795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:03:06.8515225Z logits = self.prob_head(feat_concat) 2025-09-07T08:03:06.8515354Z 2025-09-07T08:03:06.8515458Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8516033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:03:06.8516481Z logits = self.prob_head(feat_concat) 2025-09-07T08:03:06.8516605Z 2025-09-07T08:03:06.8516704Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8517157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:03:06.8517580Z logits = self.prob_head(feat_concat) 2025-09-07T08:03:06.8517708Z 2025-09-07T08:03:06.8517797Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8518247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8518681Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8519020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8519355Z x = module(x) 2025-09-07T08:03:06.8519449Z 2025-09-07T08:03:06.8519539Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8519993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8520422Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8520738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8521061Z x = module(x) 2025-09-07T08:03:06.8521147Z 2025-09-07T08:03:06.8521236Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8521688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8522119Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8522435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8522759Z x = module(x) 2025-09-07T08:03:06.8522843Z 2025-09-07T08:03:06.8522931Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8523382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8523810Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8524117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8524440Z x = module(x) 2025-09-07T08:03:06.8524738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T08:03:06.8525072Z out = self.conv1(x) 2025-09-07T08:03:06.8525167Z 2025-09-07T08:03:06.8525256Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8525785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8526211Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8526530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8526854Z x = module(x) 2025-09-07T08:03:06.8527139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:03:06.8527469Z out = self.relu(out) 2025-09-07T08:03:06.8527578Z 2025-09-07T08:03:06.8527672Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8528127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8528545Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8528933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8529264Z x = module(x) 2025-09-07T08:03:06.8529566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T08:03:06.8529900Z out = self.conv2(out) 2025-09-07T08:03:06.8530004Z 2025-09-07T08:03:06.8530096Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8530550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8530977Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8531302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8531625Z x = module(x) 2025-09-07T08:03:06.8531910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:03:06.8532259Z out = self.relu(out) 2025-09-07T08:03:06.8532362Z 2025-09-07T08:03:06.8532463Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8532929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8533356Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8533679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8534005Z x = module(x) 2025-09-07T08:03:06.8534349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T08:03:06.8534687Z out = self.conv3(out) 2025-09-07T08:03:06.8534784Z 2025-09-07T08:03:06.8534877Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8535337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8535769Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8536087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8536409Z x = module(x) 2025-09-07T08:03:06.8536691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 158, in forward 2025-09-07T08:03:06.8537033Z identity = self.downsample(x) 2025-09-07T08:03:06.8537152Z 2025-09-07T08:03:06.8537252Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8537752Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8538172Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8538492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8538808Z x = module(x) 2025-09-07T08:03:06.8539172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T08:03:06.8539494Z out = self.conv1(x) 2025-09-07T08:03:06.8539586Z 2025-09-07T08:03:06.8539676Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8540124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8540548Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8540856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8541168Z x = module(x) 2025-09-07T08:03:06.8541442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:03:06.8541771Z out = self.relu(out) 2025-09-07T08:03:06.8541863Z 2025-09-07T08:03:06.8541954Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8542472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8542907Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8543230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8543546Z x = module(x) 2025-09-07T08:03:06.8543836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T08:03:06.8544167Z out = self.conv2(out) 2025-09-07T08:03:06.8544263Z 2025-09-07T08:03:06.8544354Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8544797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8545215Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8545531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8545840Z x = module(x) 2025-09-07T08:03:06.8546121Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:03:06.8546441Z out = self.relu(out) 2025-09-07T08:03:06.8546536Z 2025-09-07T08:03:06.8546628Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8547070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8547484Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8547792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8548103Z x = module(x) 2025-09-07T08:03:06.8548384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T08:03:06.8548710Z out = self.conv3(out) 2025-09-07T08:03:06.8548806Z 2025-09-07T08:03:06.8548893Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8549332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8549748Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8550052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8550361Z x = module(x) 2025-09-07T08:03:06.8550637Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:03:06.8550956Z out = self.relu(out) 2025-09-07T08:03:06.8551047Z 2025-09-07T08:03:06.8551136Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8551577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8552078Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8552408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8552730Z x = module(x) 2025-09-07T08:03:06.8553024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T08:03:06.8553347Z out = self.conv1(x) 2025-09-07T08:03:06.8553447Z 2025-09-07T08:03:06.8553543Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8553997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8554421Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8554736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8555046Z x = module(x) 2025-09-07T08:03:06.8555390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:03:06.8555720Z out = self.relu(out) 2025-09-07T08:03:06.8555817Z 2025-09-07T08:03:06.8555910Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8556357Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8556774Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8557083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8557397Z x = module(x) 2025-09-07T08:03:06.8557676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T08:03:06.8558002Z out = self.conv2(out) 2025-09-07T08:03:06.8558104Z 2025-09-07T08:03:06.8558190Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8558638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8559055Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8559362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8559670Z x = module(x) 2025-09-07T08:03:06.8559946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:03:06.8560269Z out = self.relu(out) 2025-09-07T08:03:06.8560361Z 2025-09-07T08:03:06.8560449Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8560889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8561301Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8561618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8561932Z x = module(x) 2025-09-07T08:03:06.8562208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T08:03:06.8562527Z out = self.conv3(out) 2025-09-07T08:03:06.8562626Z 2025-09-07T08:03:06.8562712Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8563154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8563576Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8563883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8564194Z x = module(x) 2025-09-07T08:03:06.8564474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:03:06.8565338Z out = self.relu(out) 2025-09-07T08:03:06.8565431Z 2025-09-07T08:03:06.8565521Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8565965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8566387Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8566699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8567014Z x = module(x) 2025-09-07T08:03:06.8567300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T08:03:06.8567616Z out = self.conv1(x) 2025-09-07T08:03:06.8567714Z 2025-09-07T08:03:06.8567802Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8568242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8568733Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8569047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8569360Z x = module(x) 2025-09-07T08:03:06.8569642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:03:06.8569968Z out = self.relu(out) 2025-09-07T08:03:06.8570058Z 2025-09-07T08:03:06.8570150Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8570591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8571007Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8571317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8571635Z x = module(x) 2025-09-07T08:03:06.8571926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T08:03:06.8572243Z out = self.conv2(out) 2025-09-07T08:03:06.8572342Z 2025-09-07T08:03:06.8572428Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8572870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8573299Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8573601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8573912Z x = module(x) 2025-09-07T08:03:06.8574202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:03:06.8574585Z out = self.relu(out) 2025-09-07T08:03:06.8574675Z 2025-09-07T08:03:06.8574768Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8575211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8575629Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8575938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8576250Z x = module(x) 2025-09-07T08:03:06.8576532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T08:03:06.8576857Z out = self.conv3(out) 2025-09-07T08:03:06.8576957Z 2025-09-07T08:03:06.8577042Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8577491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8577952Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8578334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8578657Z x = module(x) 2025-09-07T08:03:06.8578944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 158, in forward 2025-09-07T08:03:06.8579280Z identity = self.downsample(x) 2025-09-07T08:03:06.8579401Z 2025-09-07T08:03:06.8579495Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8579939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8580363Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8580673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8580991Z x = module(x) 2025-09-07T08:03:06.8581269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T08:03:06.8581671Z out = self.conv1(x) 2025-09-07T08:03:06.8581770Z 2025-09-07T08:03:06.8581859Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8582309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8582728Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8583034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8583350Z x = module(x) 2025-09-07T08:03:06.8583634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:03:06.8583957Z out = self.relu(out) 2025-09-07T08:03:06.8584049Z 2025-09-07T08:03:06.8584140Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8584582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8585008Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8585321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8585633Z x = module(x) 2025-09-07T08:03:06.8585910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T08:03:06.8586233Z out = self.conv2(out) 2025-09-07T08:03:06.8586327Z 2025-09-07T08:03:06.8586412Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8586854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8587278Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8587581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8587904Z x = module(x) 2025-09-07T08:03:06.8588183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:03:06.8588506Z out = self.relu(out) 2025-09-07T08:03:06.8588594Z 2025-09-07T08:03:06.8588686Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8589131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8589549Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8589852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8590162Z x = module(x) 2025-09-07T08:03:06.8590436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T08:03:06.8590755Z out = self.conv3(out) 2025-09-07T08:03:06.8590847Z 2025-09-07T08:03:06.8590931Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8591514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8591930Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8592237Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8592548Z x = module(x) 2025-09-07T08:03:06.8592828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:03:06.8593147Z out = self.relu(out) 2025-09-07T08:03:06.8593241Z 2025-09-07T08:03:06.8593329Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8593770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8594185Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8594554Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8594869Z x = module(x) 2025-09-07T08:03:06.8595144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T08:03:06.8595461Z out = self.conv1(x) 2025-09-07T08:03:06.8595561Z 2025-09-07T08:03:06.8595658Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8596099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8596523Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8596834Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8597160Z x = module(x) 2025-09-07T08:03:06.8597438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:03:06.8597774Z out = self.relu(out) 2025-09-07T08:03:06.8597874Z 2025-09-07T08:03:06.8597963Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8598414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8598843Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8599154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8599475Z x = module(x) 2025-09-07T08:03:06.8599759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T08:03:06.8600089Z out = self.conv2(out) 2025-09-07T08:03:06.8600184Z 2025-09-07T08:03:06.8600276Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8600721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8601145Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8601459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8601775Z x = module(x) 2025-09-07T08:03:06.8602056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:03:06.8602378Z out = self.relu(out) 2025-09-07T08:03:06.8602475Z 2025-09-07T08:03:06.8602562Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8603005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8603572Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8603891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8604323Z x = module(x) 2025-09-07T08:03:06.8604620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T08:03:06.8604954Z out = self.conv3(out) 2025-09-07T08:03:06.8605055Z 2025-09-07T08:03:06.8605157Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8605614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8606078Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8606409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8606743Z x = module(x) 2025-09-07T08:03:06.8607037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:03:06.8607368Z out = self.relu(out) 2025-09-07T08:03:06.8607473Z 2025-09-07T08:03:06.8607568Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8608118Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8608555Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8608873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8609208Z x = module(x) 2025-09-07T08:03:06.8609503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T08:03:06.8609834Z out = self.conv1(x) 2025-09-07T08:03:06.8609924Z 2025-09-07T08:03:06.8610015Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8610468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8610893Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8611214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8611530Z x = module(x) 2025-09-07T08:03:06.8611810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:03:06.8612135Z out = self.relu(out) 2025-09-07T08:03:06.8612227Z 2025-09-07T08:03:06.8612315Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8612771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8613200Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8613512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8613833Z x = module(x) 2025-09-07T08:03:06.8614122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T08:03:06.8614459Z out = self.conv2(out) 2025-09-07T08:03:06.8614558Z 2025-09-07T08:03:06.8614646Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8615099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8615529Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8615841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8616190Z x = module(x) 2025-09-07T08:03:06.8616472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:03:06.8616801Z out = self.relu(out) 2025-09-07T08:03:06.8616900Z 2025-09-07T08:03:06.8616991Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8617444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8617989Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8618301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8618624Z x = module(x) 2025-09-07T08:03:06.8618917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T08:03:06.8619247Z out = self.conv3(out) 2025-09-07T08:03:06.8619343Z 2025-09-07T08:03:06.8619434Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8619885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8620304Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8620619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8620941Z x = module(x) 2025-09-07T08:03:06.8621288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:03:06.8621626Z out = self.relu(out) 2025-09-07T08:03:06.8621718Z 2025-09-07T08:03:06.8621813Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8622264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8622688Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8623004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8623321Z x = module(x) 2025-09-07T08:03:06.8623608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T08:03:06.8623938Z out = self.conv1(x) 2025-09-07T08:03:06.8624028Z 2025-09-07T08:03:06.8624119Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8624571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8624999Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8625316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8625639Z x = module(x) 2025-09-07T08:03:06.8625919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:03:06.8626248Z out = self.relu(out) 2025-09-07T08:03:06.8626341Z 2025-09-07T08:03:06.8626435Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8626881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8627300Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8627620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8627940Z x = module(x) 2025-09-07T08:03:06.8628231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T08:03:06.8628567Z out = self.conv2(out) 2025-09-07T08:03:06.8628662Z 2025-09-07T08:03:06.8628753Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8629204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8629625Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8629943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8630261Z x = module(x) 2025-09-07T08:03:06.8630548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:03:06.8630964Z out = self.relu(out) 2025-09-07T08:03:06.8631070Z 2025-09-07T08:03:06.8631168Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8631616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8632036Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8632353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8632672Z x = module(x) 2025-09-07T08:03:06.8632963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T08:03:06.8633291Z out = self.conv3(out) 2025-09-07T08:03:06.8633386Z 2025-09-07T08:03:06.8633477Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8633987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8634427Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8634745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8635137Z x = module(x) 2025-09-07T08:03:06.8635437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 158, in forward 2025-09-07T08:03:06.8635776Z identity = self.downsample(x) 2025-09-07T08:03:06.8635896Z 2025-09-07T08:03:06.8635995Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8636457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8636884Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8637222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8637552Z x = module(x) 2025-09-07T08:03:06.8637854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T08:03:06.8638187Z out = self.conv1(x) 2025-09-07T08:03:06.8638284Z 2025-09-07T08:03:06.8638381Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8638840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8639270Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8639593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8639911Z x = module(x) 2025-09-07T08:03:06.8640204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:03:06.8640537Z out = self.relu(out) 2025-09-07T08:03:06.8640637Z 2025-09-07T08:03:06.8640735Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8641188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8641610Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8641923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8642241Z x = module(x) 2025-09-07T08:03:06.8642530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T08:03:06.8642853Z out = self.conv2(out) 2025-09-07T08:03:06.8642959Z 2025-09-07T08:03:06.8643048Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8643495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8643931Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8644335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8644651Z x = module(x) 2025-09-07T08:03:06.8644945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:03:06.8645273Z out = self.relu(out) 2025-09-07T08:03:06.8645371Z 2025-09-07T08:03:06.8645469Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8645921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8646343Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8646661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8646983Z x = module(x) 2025-09-07T08:03:06.8647269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T08:03:06.8647659Z out = self.conv3(out) 2025-09-07T08:03:06.8647763Z 2025-09-07T08:03:06.8647854Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8648314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8648743Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8649066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8649382Z x = module(x) 2025-09-07T08:03:06.8649669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:03:06.8650026Z out = self.relu(out) 2025-09-07T08:03:06.8650120Z 2025-09-07T08:03:06.8650216Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8650673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8651095Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8651412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8651732Z x = module(x) 2025-09-07T08:03:06.8652019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T08:03:06.8652344Z out = self.conv1(x) 2025-09-07T08:03:06.8652443Z 2025-09-07T08:03:06.8652533Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8652988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8653426Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8653742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8654062Z x = module(x) 2025-09-07T08:03:06.8654352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:03:06.8654681Z out = self.relu(out) 2025-09-07T08:03:06.8654775Z 2025-09-07T08:03:06.8654871Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8655315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8655743Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8656061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8656383Z x = module(x) 2025-09-07T08:03:06.8656671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T08:03:06.8656996Z out = self.conv2(out) 2025-09-07T08:03:06.8657099Z 2025-09-07T08:03:06.8657195Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8657799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8658231Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8658552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8658871Z x = module(x) 2025-09-07T08:03:06.8659163Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:03:06.8659493Z out = self.relu(out) 2025-09-07T08:03:06.8659586Z 2025-09-07T08:03:06.8659682Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8660127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8660560Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8660947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8661278Z x = module(x) 2025-09-07T08:03:06.8661568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T08:03:06.8661895Z out = self.conv3(out) 2025-09-07T08:03:06.8661998Z 2025-09-07T08:03:06.8662088Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8662536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8662964Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8663285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8663602Z x = module(x) 2025-09-07T08:03:06.8663892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:03:06.8664233Z out = self.relu(out) 2025-09-07T08:03:06.8664325Z 2025-09-07T08:03:06.8664422Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8664868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8665296Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8665621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8665950Z x = module(x) 2025-09-07T08:03:06.8666239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T08:03:06.8666570Z out = self.conv1(x) 2025-09-07T08:03:06.8666668Z 2025-09-07T08:03:06.8666760Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8667210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8667653Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8667977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8668308Z x = module(x) 2025-09-07T08:03:06.8668603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:03:06.8668942Z out = self.relu(out) 2025-09-07T08:03:06.8669041Z 2025-09-07T08:03:06.8669142Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8669593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8670024Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8670344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8670748Z x = module(x) 2025-09-07T08:03:06.8671036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T08:03:06.8671374Z out = self.conv2(out) 2025-09-07T08:03:06.8671483Z 2025-09-07T08:03:06.8671574Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8672031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8672463Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8672776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8673103Z x = module(x) 2025-09-07T08:03:06.8673395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:03:06.8673729Z out = self.relu(out) 2025-09-07T08:03:06.8673821Z 2025-09-07T08:03:06.8673919Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8674440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8674882Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8675208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8675532Z x = module(x) 2025-09-07T08:03:06.8675819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T08:03:06.8676154Z out = self.conv3(out) 2025-09-07T08:03:06.8676257Z 2025-09-07T08:03:06.8676345Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8676797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8677226Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8677540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8677867Z x = module(x) 2025-09-07T08:03:06.8678169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:03:06.8678504Z out = self.relu(out) 2025-09-07T08:03:06.8678597Z 2025-09-07T08:03:06.8678694Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8679141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8679566Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8679889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8680213Z x = module(x) 2025-09-07T08:03:06.8680494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T08:03:06.8680833Z out = self.conv1(x) 2025-09-07T08:03:06.8680932Z 2025-09-07T08:03:06.8681022Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8681475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8681909Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8682221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8682547Z x = module(x) 2025-09-07T08:03:06.8682840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:03:06.8683173Z out = self.relu(out) 2025-09-07T08:03:06.8683264Z 2025-09-07T08:03:06.8683356Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8683814Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8684341Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8684667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8684990Z x = module(x) 2025-09-07T08:03:06.8685295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T08:03:06.8685640Z out = self.conv2(out) 2025-09-07T08:03:06.8685745Z 2025-09-07T08:03:06.8685838Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8686293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8686734Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8687044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8687372Z x = module(x) 2025-09-07T08:03:06.8687736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:03:06.8688076Z out = self.relu(out) 2025-09-07T08:03:06.8688171Z 2025-09-07T08:03:06.8688260Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8688716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8689142Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8689461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8689781Z x = module(x) 2025-09-07T08:03:06.8690066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T08:03:06.8690399Z out = self.conv3(out) 2025-09-07T08:03:06.8690501Z 2025-09-07T08:03:06.8690590Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8691047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8691477Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8691787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8692112Z x = module(x) 2025-09-07T08:03:06.8692404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:03:06.8692734Z out = self.relu(out) 2025-09-07T08:03:06.8692827Z 2025-09-07T08:03:06.8692917Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8693368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8693796Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8694120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8694442Z x = module(x) 2025-09-07T08:03:06.8694723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T08:03:06.8695054Z out = self.conv1(x) 2025-09-07T08:03:06.8695153Z 2025-09-07T08:03:06.8695246Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8695701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8696121Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8696437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8696763Z x = module(x) 2025-09-07T08:03:06.8697053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:03:06.8697462Z out = self.relu(out) 2025-09-07T08:03:06.8697555Z 2025-09-07T08:03:06.8697647Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8698142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8698570Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8698889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8699210Z x = module(x) 2025-09-07T08:03:06.8699494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T08:03:06.8699826Z out = self.conv2(out) 2025-09-07T08:03:06.8699923Z 2025-09-07T08:03:06.8700024Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8700541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8700973Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8701297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8701621Z x = module(x) 2025-09-07T08:03:06.8701909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:03:06.8702244Z out = self.relu(out) 2025-09-07T08:03:06.8702336Z 2025-09-07T08:03:06.8702428Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8702878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8703433Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8703760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8704074Z x = module(x) 2025-09-07T08:03:06.8704374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T08:03:06.8704710Z out = self.conv3(out) 2025-09-07T08:03:06.8704811Z 2025-09-07T08:03:06.8704916Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8705372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8705795Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8706114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8706439Z x = module(x) 2025-09-07T08:03:06.8706731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:03:06.8707063Z out = self.relu(out) 2025-09-07T08:03:06.8707163Z 2025-09-07T08:03:06.8707257Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8707717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8708143Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8708462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8708777Z x = module(x) 2025-09-07T08:03:06.8709072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T08:03:06.8709404Z out = self.conv1(x) 2025-09-07T08:03:06.8709498Z 2025-09-07T08:03:06.8709597Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8710051Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8710472Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8710903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8711223Z x = module(x) 2025-09-07T08:03:06.8711512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:03:06.8711841Z out = self.relu(out) 2025-09-07T08:03:06.8711945Z 2025-09-07T08:03:06.8712036Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8712487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8712918Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8713236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8713554Z x = module(x) 2025-09-07T08:03:06.8713847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T08:03:06.8714275Z out = self.conv2(out) 2025-09-07T08:03:06.8714376Z 2025-09-07T08:03:06.8714476Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8714930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8715355Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8715674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8716000Z x = module(x) 2025-09-07T08:03:06.8716291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:03:06.8716616Z out = self.relu(out) 2025-09-07T08:03:06.8716717Z 2025-09-07T08:03:06.8716809Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8717290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8717739Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8718085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8718415Z x = module(x) 2025-09-07T08:03:06.8718720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T08:03:06.8719058Z out = self.conv3(out) 2025-09-07T08:03:06.8719161Z 2025-09-07T08:03:06.8719269Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8719729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8720154Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8720479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8720796Z x = module(x) 2025-09-07T08:03:06.8721098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 158, in forward 2025-09-07T08:03:06.8721437Z identity = self.downsample(x) 2025-09-07T08:03:06.8721564Z 2025-09-07T08:03:06.8721657Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8722112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8722545Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8722860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8723177Z x = module(x) 2025-09-07T08:03:06.8723484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T08:03:06.8723821Z out = self.conv1(x) 2025-09-07T08:03:06.8723920Z 2025-09-07T08:03:06.8724017Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8724550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8724977Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8725300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8725624Z x = module(x) 2025-09-07T08:03:06.8725935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:03:06.8726261Z out = self.relu(out) 2025-09-07T08:03:06.8726371Z 2025-09-07T08:03:06.8726466Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8726921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8727352Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8727738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8728060Z x = module(x) 2025-09-07T08:03:06.8728357Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T08:03:06.8728692Z out = self.conv2(out) 2025-09-07T08:03:06.8728794Z 2025-09-07T08:03:06.8728898Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8729347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8729786Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8730105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8730428Z x = module(x) 2025-09-07T08:03:06.8739096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:03:06.8739529Z out = self.relu(out) 2025-09-07T08:03:06.8739652Z 2025-09-07T08:03:06.8739760Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8740272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8740752Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8741112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8741445Z x = module(x) 2025-09-07T08:03:06.8741758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T08:03:06.8742103Z out = self.conv3(out) 2025-09-07T08:03:06.8742208Z 2025-09-07T08:03:06.8742316Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8742786Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8743224Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8743556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8743874Z x = module(x) 2025-09-07T08:03:06.8744166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:03:06.8744490Z out = self.relu(out) 2025-09-07T08:03:06.8744596Z 2025-09-07T08:03:06.8744693Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8745149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8745575Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8745899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8746331Z x = module(x) 2025-09-07T08:03:06.8746633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T08:03:06.8746967Z out = self.conv1(x) 2025-09-07T08:03:06.8747064Z 2025-09-07T08:03:06.8747165Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8747622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8748064Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8748393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8748712Z x = module(x) 2025-09-07T08:03:06.8749000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T08:03:06.8749324Z out = self.relu(out) 2025-09-07T08:03:06.8749427Z 2025-09-07T08:03:06.8749522Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8750058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8750494Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8750807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8751130Z x = module(x) 2025-09-07T08:03:06.8751420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T08:03:06.8751754Z out = self.conv2(out) 2025-09-07T08:03:06.8751853Z 2025-09-07T08:03:06.8751948Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8752397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8752818Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8753139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8753456Z x = module(x) 2025-09-07T08:03:06.8753738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T08:03:06.8754066Z out = self.relu(out) 2025-09-07T08:03:06.8754166Z 2025-09-07T08:03:06.8754255Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8754703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8755141Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8755447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8755767Z x = module(x) 2025-09-07T08:03:06.8756086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T08:03:06.8756419Z out = self.conv3(out) 2025-09-07T08:03:06.8756514Z 2025-09-07T08:03:06.8756607Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8757050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T08:03:06.8757471Z feats = self.feat_extractor(x) 2025-09-07T08:03:06.8757781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T08:03:06.8758097Z x = module(x) 2025-09-07T08:03:06.8758373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T08:03:06.8758712Z out = self.relu(out) 2025-09-07T08:03:06.8758811Z 2025-09-07T08:03:06.8758902Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8759358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8759880Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8760297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:03:06.8760803Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:03:06.8761316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:03:06.8761813Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:03:06.8762003Z 2025-09-07T08:03:06.8762103Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8762548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8762981Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8763470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:03:06.8763967Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:03:06.8764464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:03:06.8764951Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:03:06.8765142Z 2025-09-07T08:03:06.8765236Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8765689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8766115Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8766525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:03:06.8767013Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:03:06.8767503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:03:06.8767992Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:03:06.8768176Z 2025-09-07T08:03:06.8768276Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8768740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8769169Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8769590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 86, in forward 2025-09-07T08:03:06.8770043Z out.append(self.upsample(out[-1]) + t) 2025-09-07T08:03:06.8770178Z 2025-09-07T08:03:06.8770279Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8770727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8771149Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8771561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:03:06.8772048Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:03:06.8772547Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:03:06.8773036Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:03:06.8773222Z 2025-09-07T08:03:06.8773392Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8773841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8774258Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8774662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 86, in forward 2025-09-07T08:03:06.8775091Z out.append(self.upsample(out[-1]) + t) 2025-09-07T08:03:06.8775219Z 2025-09-07T08:03:06.8775308Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8775753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8776177Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8776576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T08:03:06.8777120Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:03:06.8777610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T08:03:06.8778156Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T08:03:06.8778347Z 2025-09-07T08:03:06.8778444Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8778900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8779327Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8779737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 86, in forward 2025-09-07T08:03:06.8780168Z out.append(self.upsample(out[-1]) + t) 2025-09-07T08:03:06.8780306Z 2025-09-07T08:03:06.8780400Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8780847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8781265Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8781667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:03:06.8782170Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8782666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:03:06.8783161Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8783329Z 2025-09-07T08:03:06.8783424Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8783876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8784299Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8784703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:03:06.8785168Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8785638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:03:06.8786106Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8786275Z 2025-09-07T08:03:06.8786373Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8786826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8787329Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8787736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:03:06.8788203Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8788668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:03:06.8789140Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8789307Z 2025-09-07T08:03:06.8789398Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8789838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8790256Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8790722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:03:06.8791189Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8791659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:03:06.8792129Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8792293Z 2025-09-07T08:03:06.8792387Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8792858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8793276Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8793679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:03:06.8794145Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8794611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:03:06.8795077Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8795242Z 2025-09-07T08:03:06.8795331Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8795770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8796189Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8796589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:03:06.8797045Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8797516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:03:06.8797982Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8798157Z 2025-09-07T08:03:06.8798246Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8798690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8799108Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8799507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:03:06.8799966Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8800431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:03:06.8800993Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8801165Z 2025-09-07T08:03:06.8801255Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8801698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8802117Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8802513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:03:06.8804955Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8805497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:03:06.8806251Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8806433Z 2025-09-07T08:03:06.8806540Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8807003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8807434Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8807850Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:03:06.8808325Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8808802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:03:06.8809270Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8809439Z 2025-09-07T08:03:06.8809532Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8810004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8810443Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8810856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:03:06.8811335Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8811817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:03:06.8812291Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8812458Z 2025-09-07T08:03:06.8812556Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8813011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8813435Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8813842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:03:06.8814306Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8814777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:03:06.8815253Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8815418Z 2025-09-07T08:03:06.8815509Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8815951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8816483Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8816901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T08:03:06.8817378Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8817910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T08:03:06.8818383Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T08:03:06.8818553Z 2025-09-07T08:03:06.8818646Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8819097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T08:03:06.8819518Z feat_concat = self.fpn(feats) 2025-09-07T08:03:06.8819986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 91, in forward 2025-09-07T08:03:06.8820416Z return torch.cat(out, dim=1) 2025-09-07T08:03:06.8820532Z 2025-09-07T08:03:06.8820608Z cudagraph partition due to non gpu ops 2025-09-07T08:03:06.8820833Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8821282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:03:06.8821708Z logits = self.prob_head(feat_concat) 2025-09-07T08:03:06.8821838Z 2025-09-07T08:03:06.8821930Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8822376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:03:06.8822799Z logits = self.prob_head(feat_concat) 2025-09-07T08:03:06.8822920Z 2025-09-07T08:03:06.8823012Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8823455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:03:06.8823877Z logits = self.prob_head(feat_concat) 2025-09-07T08:03:06.8823998Z 2025-09-07T08:03:06.8824085Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8824523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T08:03:06.8824947Z logits = self.prob_head(feat_concat) 2025-09-07T08:03:06.8825063Z 2025-09-07T08:03:06.8825133Z cudagraph partition due to non gpu ops 2025-09-07T08:03:06.8825345Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:06.8825803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 202, in forward 2025-09-07T08:03:06.8826252Z prob_map = _bf16_to_float32(torch.sigmoid(logits)) 2025-09-07T08:03:06.8826641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/utils/pytorch.py", line 30, in _bf16_to_float32 2025-09-07T08:03:06.8827028Z return x.float() if x.dtype == torch.bfloat16 else x 2025-09-07T08:03:06.8827178Z 2025-09-07T08:03:40.9132332Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:40.9132853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/core.py", line 95, in __call__ 2025-09-07T08:03:40.9133302Z for bmap in (proba_map >= self.bin_thresh).astype(np.uint8) 2025-09-07T08:03:40.9133502Z 2025-09-07T08:03:41.6699624Z pass 2025-09-07T08:03:41.6703296Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:03:44.5842662Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:03:44.5844097Z import pynvml # type: ignore[import] 2025-09-07T08:03:46.5809343Z 2025-09-07T08:03:49.1854356Z loading model: 0it [00:00, ?it/s]WARNING:common:Model doctr_reco_predictor does not support bfloat16, running with amp instead 2025-09-07T08:03:49.2282881Z 2025-09-07T08:03:49.2283493Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:03:49.2283929Z WARNING:common:Model doctr_reco_predictor does not support bfloat16, running with amp instead 2025-09-07T08:03:49.2285460Z cpu eval doctr_reco_predictor 2025-09-07T08:03:49.7717396Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:03:49.7717974Z WARNING:common:Model doctr_reco_predictor does not support bfloat16, running with amp instead 2025-09-07T08:03:49.8322440Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:03:49.8785153Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:03:58.5438771Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5439312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5439752Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5439891Z 2025-09-07T08:03:58.5440009Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5440457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5440861Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5441012Z 2025-09-07T08:03:58.5441113Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5441529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5441962Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5442081Z 2025-09-07T08:03:58.5442185Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5442622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5443001Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5443117Z 2025-09-07T08:03:58.5443211Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5443612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5444003Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5444130Z 2025-09-07T08:03:58.5444233Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5444637Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5445016Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5445145Z 2025-09-07T08:03:58.5445239Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5445632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5446009Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5446128Z 2025-09-07T08:03:58.5446224Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5446613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5446990Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5447111Z 2025-09-07T08:03:58.5447200Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5447597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5447966Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5448501Z 2025-09-07T08:03:58.5448598Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5448995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5449367Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5449482Z 2025-09-07T08:03:58.5449577Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5449961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5450333Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5450456Z 2025-09-07T08:03:58.5450544Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5450936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5451309Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5451429Z 2025-09-07T08:03:58.5451648Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5452043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5452412Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5452531Z 2025-09-07T08:03:58.5452630Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5453021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5453389Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5453511Z 2025-09-07T08:03:58.5453602Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5453998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5454367Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5454484Z 2025-09-07T08:03:58.5454583Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5454971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5455341Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5455464Z 2025-09-07T08:03:58.5455555Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5455948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5456317Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5456435Z 2025-09-07T08:03:58.5456523Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5456915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5457284Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5457401Z 2025-09-07T08:03:58.5457503Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5457948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5458324Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5458464Z 2025-09-07T08:03:58.5458556Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5458948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5459319Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5459434Z 2025-09-07T08:03:58.5459523Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5459910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5460282Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5460397Z 2025-09-07T08:03:58.5460493Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5460974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5461337Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5461461Z 2025-09-07T08:03:58.5461550Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5461938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5462312Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5462428Z 2025-09-07T08:03:58.5462525Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5462911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5463286Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5463407Z 2025-09-07T08:03:58.5463496Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5463955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5464335Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5464452Z 2025-09-07T08:03:58.5464542Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5464929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5465300Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5465415Z 2025-09-07T08:03:58.5465508Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5465897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5466264Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5466384Z 2025-09-07T08:03:58.5466471Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5466875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5467241Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5467357Z 2025-09-07T08:03:58.5467447Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5467840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5468209Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5468324Z 2025-09-07T08:03:58.5468421Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5468805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5469171Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5469293Z 2025-09-07T08:03:58.5469379Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5469776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5470148Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5470263Z 2025-09-07T08:03:58.5470360Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5470746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5471120Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5471240Z 2025-09-07T08:03:58.5471330Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5471717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5472092Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5472210Z 2025-09-07T08:03:58.5472299Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5472706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5473165Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5473280Z 2025-09-07T08:03:58.5473380Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5473769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5474149Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5474277Z 2025-09-07T08:03:58.5474366Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5474762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5475133Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5475249Z 2025-09-07T08:03:58.5475339Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5475795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5476178Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5476294Z 2025-09-07T08:03:58.5476392Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5476783Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5477150Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5477274Z 2025-09-07T08:03:58.5477363Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5477753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5478127Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5478244Z 2025-09-07T08:03:58.5478340Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5478739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5479124Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5479244Z 2025-09-07T08:03:58.5479333Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5479724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5480096Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5480213Z 2025-09-07T08:03:58.5480303Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5480692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5481063Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5481179Z 2025-09-07T08:03:58.5481277Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5481660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5482038Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5482158Z 2025-09-07T08:03:58.5482270Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5482658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5483031Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5483148Z 2025-09-07T08:03:58.5483245Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5483629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5484001Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5484124Z 2025-09-07T08:03:58.5484214Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5484604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5485063Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5485181Z 2025-09-07T08:03:58.5485274Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5485667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5486041Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5486156Z 2025-09-07T08:03:58.5486257Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5486643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5487009Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5487135Z 2025-09-07T08:03:58.5487227Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5487622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5488013Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5488221Z 2025-09-07T08:03:58.5488316Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5488714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5489083Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5489201Z 2025-09-07T08:03:58.5489300Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5489690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5490054Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5490177Z 2025-09-07T08:03:58.5490267Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5490660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5491029Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5491146Z 2025-09-07T08:03:58.5491245Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5491629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5491998Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5492121Z 2025-09-07T08:03:58.5492211Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5492602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5492971Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5493089Z 2025-09-07T08:03:58.5493179Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5493565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5493934Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5494053Z 2025-09-07T08:03:58.5494154Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5494547Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5494915Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5495037Z 2025-09-07T08:03:58.5495127Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5495516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5495891Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5496007Z 2025-09-07T08:03:58.5496099Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:03:58.5496491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 206, in forward 2025-09-07T08:03:58.5496867Z features = self.feat_extractor(x) 2025-09-07T08:03:58.5496982Z 2025-09-07T08:04:07.6481452Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:07.6481997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 75, in ctc_best_path 2025-09-07T08:04:07.6482508Z probs = F.softmax(logits, dim=-1).max(dim=-1).values.min(dim=1).values 2025-09-07T08:04:07.6482709Z 2025-09-07T08:04:07.8450419Z pass 2025-09-07T08:04:07.8454672Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:10.0531609Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:04:10.0532449Z import pynvml # type: ignore[import] 2025-09-07T08:04:12.0572511Z 2025-09-07T08:04:12.1101900Z loading model: 0it [00:00, ?it/s]Gym has been unmaintained since 2022 and does not support NumPy 2.0 amongst other critical functionality. 2025-09-07T08:04:12.1102597Z Please upgrade to Gymnasium, the maintained drop-in replacement of Gym, or contact the authors of your software and request that they upgrade. 2025-09-07T08:04:12.1103373Z Users of this version of Gym should be able to simply replace 'import gym' with 'import gymnasium as gym' in the vast majority of cases. 2025-09-07T08:04:12.1103971Z See the migration guide at https://gymnasium.farama.org/introduction/migration_guide/ for additional information. 2025-09-07T08:04:13.0011944Z 2025-09-07T08:04:13.0012298Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:04:13.0026510Z cpu eval drq 2025-09-07T08:04:13.0204981Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:13.0311008Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:13.0388586Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:15.0362267Z cudagraph partition due to non gpu ops 2025-09-07T08:04:15.0362561Z cudagraph partition due to non gpu ops 2025-09-07T08:04:15.0362783Z cudagraph partition due to non gpu ops 2025-09-07T08:04:15.0362976Z cudagraph partition due to non gpu ops 2025-09-07T08:04:15.0363216Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:15.0363602Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:15.0363949Z return mod(*inputs) 2025-09-07T08:04:15.0364211Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:04:15.0364522Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:04:15.0364817Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:04:15.0365092Z h = self.forward_conv(obs) 2025-09-07T08:04:15.0365389Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 43, in forward_conv 2025-09-07T08:04:15.0365693Z conv = torch.relu(self.convs[0](obs)) 2025-09-07T08:04:15.0365830Z 2025-09-07T08:04:15.0365935Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:15.0366300Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:15.0366624Z return mod(*inputs) 2025-09-07T08:04:15.0366853Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:04:15.0367140Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:04:15.0367411Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:04:15.0367672Z h = self.forward_conv(obs) 2025-09-07T08:04:15.0367928Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 47, in forward_conv 2025-09-07T08:04:15.0368211Z conv = torch.relu(self.convs[i](conv)) 2025-09-07T08:04:15.0368340Z 2025-09-07T08:04:15.0370258Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:15.0370650Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:15.0370985Z return mod(*inputs) 2025-09-07T08:04:15.0371220Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:04:15.0371509Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:04:15.0371787Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:04:15.0372049Z h = self.forward_conv(obs) 2025-09-07T08:04:15.0372301Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 47, in forward_conv 2025-09-07T08:04:15.0372590Z conv = torch.relu(self.convs[i](conv)) 2025-09-07T08:04:15.0372724Z 2025-09-07T08:04:15.0372845Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:15.0373207Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:15.0373529Z return mod(*inputs) 2025-09-07T08:04:15.0373914Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:04:15.0374192Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:04:15.0374467Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:04:15.0374746Z h = self.forward_conv(obs) 2025-09-07T08:04:15.0375003Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 47, in forward_conv 2025-09-07T08:04:15.0375283Z conv = torch.relu(self.convs[i](conv)) 2025-09-07T08:04:15.0375411Z 2025-09-07T08:04:15.0375511Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:15.0375861Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:15.0376175Z return mod(*inputs) 2025-09-07T08:04:15.0376414Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:04:15.0376701Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:04:15.0376992Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 56, in forward 2025-09-07T08:04:15.0377247Z h = self.forward_conv(obs) 2025-09-07T08:04:15.0377515Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 47, in forward_conv 2025-09-07T08:04:15.0377873Z conv = torch.relu(self.convs[i](conv)) 2025-09-07T08:04:15.0377999Z 2025-09-07T08:04:15.0378102Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:15.0378448Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:15.0378768Z return mod(*inputs) 2025-09-07T08:04:15.0378996Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:04:15.0379275Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:04:15.0379546Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 61, in forward 2025-09-07T08:04:15.0379793Z out = self.head(h) 2025-09-07T08:04:15.0379896Z 2025-09-07T08:04:15.0379999Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:15.0380350Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:15.0380666Z return mod(*inputs) 2025-09-07T08:04:15.0380897Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 97, in forward 2025-09-07T08:04:15.0381189Z obs = self.encoder(obs, detach=detach_encoder) 2025-09-07T08:04:15.0381476Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 63, in forward 2025-09-07T08:04:15.0381743Z out = torch.tanh(out) 2025-09-07T08:04:15.0381847Z 2025-09-07T08:04:15.0381950Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:15.0382292Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:15.0382609Z return mod(*inputs) 2025-09-07T08:04:15.0382847Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:04:15.0383133Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:04:15.0383362Z 2025-09-07T08:04:15.0383458Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:15.0383803Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:15.0384117Z return mod(*inputs) 2025-09-07T08:04:15.0384350Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:04:15.0384629Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:04:15.0384766Z 2025-09-07T08:04:15.0384859Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:15.0385202Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:15.0385519Z return mod(*inputs) 2025-09-07T08:04:15.0385753Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:04:15.0386026Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:04:15.0386171Z 2025-09-07T08:04:15.0386332Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:15.0386681Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:15.0386998Z return mod(*inputs) 2025-09-07T08:04:15.0387228Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:04:15.0387497Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:04:15.0387638Z 2025-09-07T08:04:15.0387730Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:15.0388069Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:15.0388396Z return mod(*inputs) 2025-09-07T08:04:15.0388637Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 99, in forward 2025-09-07T08:04:15.0388930Z mu, log_std = self.trunk(obs).chunk(2, dim=-1) 2025-09-07T08:04:15.0389082Z 2025-09-07T08:04:15.0389185Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:15.0389544Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:15.0389863Z return mod(*inputs) 2025-09-07T08:04:15.0390098Z File "/torchbench/torchbenchmark/models/drq/drq.py", line 105, in forward 2025-09-07T08:04:15.0390365Z std = log_std.exp() 2025-09-07T08:04:15.0390457Z 2025-09-07T08:04:22.6632515Z pass 2025-09-07T08:04:22.6635152Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:24.6637802Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:04:24.6638777Z import pynvml # type: ignore[import] 2025-09-07T08:04:26.6769725Z 2025-09-07T08:04:28.2404632Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:04:28.2405127Z loading model: 0it [00:01, ?it/s] 2025-09-07T08:04:28.2595646Z cpu eval fastNLP_Bert 2025-09-07T08:04:29.2642669Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:29.6611759Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:30.1199397Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:30.2478524Z W0907 08:04:30.247199 41954 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] Graph break from `Tensor.item()`, consider setting: 2025-09-07T08:04:30.2479192Z W0907 08:04:30.247199 41954 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] torch._dynamo.config.capture_scalar_outputs = True 2025-09-07T08:04:30.2479700Z W0907 08:04:30.247199 41954 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] or: 2025-09-07T08:04:30.2480225Z W0907 08:04:30.247199 41954 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1 2025-09-07T08:04:30.2481212Z W0907 08:04:30.247199 41954 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] to include these operations in the captured graph. 2025-09-07T08:04:30.2481725Z W0907 08:04:30.247199 41954 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:04:30.2482179Z W0907 08:04:30.247199 41954 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] Graph break: from user code at: 2025-09-07T08:04:30.2482842Z W0907 08:04:30.247199 41954 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:30.2483517Z W0907 08:04:30.247199 41954 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] return mod(*inputs) 2025-09-07T08:04:30.2484328Z W0907 08:04:30.247199 41954 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/models/bert.py", line 265, in forward 2025-09-07T08:04:30.2485008Z W0907 08:04:30.247199 41954 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] sequence_output = self.bert(words) 2025-09-07T08:04:30.2485718Z W0907 08:04:30.247199 41954 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 137, in forward 2025-09-07T08:04:30.2486443Z W0907 08:04:30.247199 41954 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] outputs = self.model(words) 2025-09-07T08:04:30.2487124Z W0907 08:04:30.247199 41954 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 445, in forward 2025-09-07T08:04:30.2490991Z W0907 08:04:30.247199 41954 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] max_word_piece_length = batch_word_pieces_length.sum(dim=-1).max().item() # 表示word piece的长度(包括padding) 2025-09-07T08:04:30.2491622Z W0907 08:04:30.247199 41954 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:04:30.2492003Z W0907 08:04:30.247199 41954 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:04:37.1479308Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.1479818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 445, in forward 2025-09-07T08:04:37.1480759Z max_word_piece_length = batch_word_pieces_length.sum(dim=-1).max().item() # 表示word piece的长度(包括padding) 2025-09-07T08:04:37.1481039Z 2025-09-07T08:04:37.1481151Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:37.1481589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 441, in forward 2025-09-07T08:04:37.1481974Z seq_len = word_mask.sum(dim=-1) 2025-09-07T08:04:37.1482147Z 2025-09-07T08:04:42.0582240Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0582742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 512, in forward 2025-09-07T08:04:42.0583180Z embedding_output = self.embeddings(input_ids, token_type_ids) 2025-09-07T08:04:42.0583605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 243, in forward 2025-09-07T08:04:42.0583991Z embeddings = self.LayerNorm(embeddings) 2025-09-07T08:04:42.0584138Z 2025-09-07T08:04:42.0584243Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0584636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0585019Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0585394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0586258Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0586682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0587121Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0587540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0587924Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0588305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 271, in forward 2025-09-07T08:04:42.0588675Z mixed_query_layer = self.query(hidden_states) 2025-09-07T08:04:42.0588823Z 2025-09-07T08:04:42.0588934Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0589316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0589819Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0590198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0590596Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0590991Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0591394Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0591786Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0592159Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0592530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 272, in forward 2025-09-07T08:04:42.0592912Z mixed_key_layer = self.key(hidden_states) 2025-09-07T08:04:42.0593053Z 2025-09-07T08:04:42.0593160Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0593525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0593889Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0594266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0594657Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0595031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0595432Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0595830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0596198Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0596574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 273, in forward 2025-09-07T08:04:42.0596934Z mixed_value_layer = self.value(hidden_states) 2025-09-07T08:04:42.0597081Z 2025-09-07T08:04:42.0597158Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0597363Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0597669Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0598047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0598410Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0598769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0599154Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0599534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0600011Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0600410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 321, in forward 2025-09-07T08:04:42.0600802Z attention_output = self.output(self_output, input_tensor) 2025-09-07T08:04:42.0601185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 307, in forward 2025-09-07T08:04:42.0601552Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0601684Z 2025-09-07T08:04:42.0601782Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0602144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0602511Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0602933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0603479Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0603864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0604282Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0604674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 335, in forward 2025-09-07T08:04:42.0605040Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0605174Z 2025-09-07T08:04:42.0605278Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0605664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0606042Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0606428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0606829Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0607210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0607608Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0608000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:04:42.0608390Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:42.0608757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:04:42.0609129Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:04:42.0609288Z 2025-09-07T08:04:42.0609388Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0609774Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0610150Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0610525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0610930Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0611316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 364, in forward 2025-09-07T08:04:42.0611726Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:04:42.0612132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 348, in forward 2025-09-07T08:04:42.0612483Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0612625Z 2025-09-07T08:04:42.0612722Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0613220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0613589Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0613946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0614332Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0614718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0615126Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0615530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0615906Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0616285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 271, in forward 2025-09-07T08:04:42.0616748Z mixed_query_layer = self.query(hidden_states) 2025-09-07T08:04:42.0616897Z 2025-09-07T08:04:42.0617005Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0617381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0617799Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0618158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0618548Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0629524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0630038Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0630472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0630878Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0631252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 272, in forward 2025-09-07T08:04:42.0631619Z mixed_key_layer = self.key(hidden_states) 2025-09-07T08:04:42.0631769Z 2025-09-07T08:04:42.0631875Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0632277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0632662Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0633025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0633422Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0633824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0634250Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0634655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0635036Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0635415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 273, in forward 2025-09-07T08:04:42.0635783Z mixed_value_layer = self.value(hidden_states) 2025-09-07T08:04:42.0635928Z 2025-09-07T08:04:42.0636016Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0636221Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0636438Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0636819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0637192Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0637668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0638050Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0638438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0638842Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0639249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 321, in forward 2025-09-07T08:04:42.0639648Z attention_output = self.output(self_output, input_tensor) 2025-09-07T08:04:42.0640040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 307, in forward 2025-09-07T08:04:42.0640411Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0640560Z 2025-09-07T08:04:42.0640727Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0641121Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0641498Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0641861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0642253Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0642642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0643042Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0643422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 335, in forward 2025-09-07T08:04:42.0643780Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0643917Z 2025-09-07T08:04:42.0644016Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0644391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0644761Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0645113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0645498Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0645877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0646273Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0646654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:04:42.0647034Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:42.0647420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:04:42.0647795Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:04:42.0647945Z 2025-09-07T08:04:42.0648047Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0648426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0648785Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0649145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0649527Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0649911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 364, in forward 2025-09-07T08:04:42.0650320Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:04:42.0650810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 348, in forward 2025-09-07T08:04:42.0651178Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0651315Z 2025-09-07T08:04:42.0651424Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0651810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0652198Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0652577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0652978Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0653372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0653782Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0654244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0654634Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0655016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 271, in forward 2025-09-07T08:04:42.0655385Z mixed_query_layer = self.query(hidden_states) 2025-09-07T08:04:42.0655532Z 2025-09-07T08:04:42.0655640Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0656010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0656378Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0656743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0657128Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0657508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0658007Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0658413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0658791Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0659152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 272, in forward 2025-09-07T08:04:42.0659515Z mixed_key_layer = self.key(hidden_states) 2025-09-07T08:04:42.0659649Z 2025-09-07T08:04:42.0659752Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0660123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0660484Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0660852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0661233Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0661614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0662004Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0662400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0662773Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0663140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 273, in forward 2025-09-07T08:04:42.0663499Z mixed_value_layer = self.value(hidden_states) 2025-09-07T08:04:42.0663638Z 2025-09-07T08:04:42.0663715Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0663995Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0664216Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0664584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0664946Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0665295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0665676Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0666056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0666452Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0666839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 321, in forward 2025-09-07T08:04:42.0667319Z attention_output = self.output(self_output, input_tensor) 2025-09-07T08:04:42.0667711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 307, in forward 2025-09-07T08:04:42.0668070Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0668205Z 2025-09-07T08:04:42.0668308Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0668678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0669042Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0669403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0669785Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0670162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0670548Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0670932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 335, in forward 2025-09-07T08:04:42.0671285Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0671415Z 2025-09-07T08:04:42.0671515Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0671887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0672247Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0672605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0672985Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0673364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0673750Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0674133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:04:42.0674517Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:42.0674894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:04:42.0675263Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:04:42.0675409Z 2025-09-07T08:04:42.0675502Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0675871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0676236Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0676593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0677048Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0677420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 364, in forward 2025-09-07T08:04:42.0677828Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:04:42.0678230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 348, in forward 2025-09-07T08:04:42.0678592Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0678723Z 2025-09-07T08:04:42.0678821Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0679187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0679552Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0679915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0680379Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0680757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0681161Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0681559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0681933Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0682308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 271, in forward 2025-09-07T08:04:42.0682666Z mixed_query_layer = self.query(hidden_states) 2025-09-07T08:04:42.0682819Z 2025-09-07T08:04:42.0682913Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0683280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0683652Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0684012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0684391Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0684767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0685167Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0685564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0685939Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0686314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 272, in forward 2025-09-07T08:04:42.0686666Z mixed_key_layer = self.key(hidden_states) 2025-09-07T08:04:42.0686802Z 2025-09-07T08:04:42.0686896Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0687263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0687623Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0687981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0688358Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0688737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0689131Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0689523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0689893Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0690331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 273, in forward 2025-09-07T08:04:42.0690693Z mixed_value_layer = self.value(hidden_states) 2025-09-07T08:04:42.0690838Z 2025-09-07T08:04:42.0690919Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0691123Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0691334Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0691699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0692065Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0692425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0692806Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0693243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0693647Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0694042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 321, in forward 2025-09-07T08:04:42.0694437Z attention_output = self.output(self_output, input_tensor) 2025-09-07T08:04:42.0694837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 307, in forward 2025-09-07T08:04:42.0695202Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0695344Z 2025-09-07T08:04:42.0695440Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0695811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0696178Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0696542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0696924Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0697304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0697741Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0698126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 335, in forward 2025-09-07T08:04:42.0698478Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0698606Z 2025-09-07T08:04:42.0698697Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0699064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0699427Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0699788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0700166Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0700544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0700926Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0701305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:04:42.0701687Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:42.0702057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:04:42.0702423Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:04:42.0702576Z 2025-09-07T08:04:42.0702666Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0703227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0704939Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0705293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0705684Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0706066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 364, in forward 2025-09-07T08:04:42.0706481Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:04:42.0706893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 348, in forward 2025-09-07T08:04:42.0707259Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0707405Z 2025-09-07T08:04:42.0707508Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0707991Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0708379Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0708745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0709137Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0709524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0709936Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0710338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0710712Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0711085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 271, in forward 2025-09-07T08:04:42.0711457Z mixed_query_layer = self.query(hidden_states) 2025-09-07T08:04:42.0711603Z 2025-09-07T08:04:42.0711705Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0712081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0712444Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0712807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0713204Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0713588Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0713991Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0714475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0714858Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0715231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 272, in forward 2025-09-07T08:04:42.0715589Z mixed_key_layer = self.key(hidden_states) 2025-09-07T08:04:42.0715724Z 2025-09-07T08:04:42.0715821Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0716189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0716558Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0716920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0717302Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0717677Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0718156Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0718552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0718923Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0719289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 273, in forward 2025-09-07T08:04:42.0719646Z mixed_value_layer = self.value(hidden_states) 2025-09-07T08:04:42.0719798Z 2025-09-07T08:04:42.0719876Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0720079Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0720297Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0720661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0721026Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0721447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0721838Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0722222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0722616Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0723012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 321, in forward 2025-09-07T08:04:42.0723404Z attention_output = self.output(self_output, input_tensor) 2025-09-07T08:04:42.0723802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 307, in forward 2025-09-07T08:04:42.0724162Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0724293Z 2025-09-07T08:04:42.0724389Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0724770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0725134Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0725493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0725877Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0726246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0726635Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0727008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 335, in forward 2025-09-07T08:04:42.0727357Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0727484Z 2025-09-07T08:04:42.0727573Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0727944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0728298Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0728652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0729029Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0729427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0729833Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0730211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:04:42.0730597Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:42.0730971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:04:42.0731429Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:04:42.0731580Z 2025-09-07T08:04:42.0731670Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0732037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0732402Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0732773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0733146Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0733516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 364, in forward 2025-09-07T08:04:42.0733917Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:04:42.0734381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 348, in forward 2025-09-07T08:04:42.0734730Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0734867Z 2025-09-07T08:04:42.0734963Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0735349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0735706Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0736057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0736425Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0736794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0737191Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0737721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0738099Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0738462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 271, in forward 2025-09-07T08:04:42.0738820Z mixed_query_layer = self.query(hidden_states) 2025-09-07T08:04:42.0738970Z 2025-09-07T08:04:42.0739064Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0739441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0739801Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0740151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0740525Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0740901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0741295Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0741684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0742044Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0742419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 272, in forward 2025-09-07T08:04:42.0742765Z mixed_key_layer = self.key(hidden_states) 2025-09-07T08:04:42.0742893Z 2025-09-07T08:04:42.0742986Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0743344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0743700Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0744133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0744524Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0744895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0745287Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0745677Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0746040Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0746402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 273, in forward 2025-09-07T08:04:42.0746759Z mixed_value_layer = self.value(hidden_states) 2025-09-07T08:04:42.0746901Z 2025-09-07T08:04:42.0746973Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0747233Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0747442Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0747817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0748167Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0748515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0748889Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0749259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0749641Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0750021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 321, in forward 2025-09-07T08:04:42.0750408Z attention_output = self.output(self_output, input_tensor) 2025-09-07T08:04:42.0750784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 307, in forward 2025-09-07T08:04:42.0751126Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0751255Z 2025-09-07T08:04:42.0751359Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0751719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0752069Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0752424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0752827Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0753218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0753597Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0753988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 335, in forward 2025-09-07T08:04:42.0754342Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0754468Z 2025-09-07T08:04:42.0754566Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0754925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0755290Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0755646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0756043Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0756435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0756825Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0757290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:04:42.0757671Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:42.0758050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:04:42.0758420Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:04:42.0758569Z 2025-09-07T08:04:42.0758665Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0759040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0759406Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0759768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0760221Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0760598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 364, in forward 2025-09-07T08:04:42.0761000Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:04:42.0761398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 348, in forward 2025-09-07T08:04:42.0761752Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0761882Z 2025-09-07T08:04:42.0761975Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0762342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0762704Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0763062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0763451Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0763818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0764224Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0764621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0764999Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0765370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 271, in forward 2025-09-07T08:04:42.0765726Z mixed_query_layer = self.query(hidden_states) 2025-09-07T08:04:42.0765873Z 2025-09-07T08:04:42.0765975Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0766369Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0766737Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0767093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0767469Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0767843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0768244Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0768636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0769003Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0769373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 272, in forward 2025-09-07T08:04:42.0769731Z mixed_key_layer = self.key(hidden_states) 2025-09-07T08:04:42.0769928Z 2025-09-07T08:04:42.0770029Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0770395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0770750Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0771110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0771492Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0771873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0772273Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0772657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0773023Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0773453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 273, in forward 2025-09-07T08:04:42.0773831Z mixed_value_layer = self.value(hidden_states) 2025-09-07T08:04:42.0773971Z 2025-09-07T08:04:42.0774052Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0774247Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0774465Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0774834Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0775203Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0775562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0775946Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0776326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0776727Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0777120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 321, in forward 2025-09-07T08:04:42.0777501Z attention_output = self.output(self_output, input_tensor) 2025-09-07T08:04:42.0777974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 307, in forward 2025-09-07T08:04:42.0778330Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0778461Z 2025-09-07T08:04:42.0778562Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0778930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0779286Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0779651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0780035Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0780424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0780811Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0781197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 335, in forward 2025-09-07T08:04:42.0781552Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0781683Z 2025-09-07T08:04:42.0781782Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0782154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0782511Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0782870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0783342Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0783720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0784108Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0784482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:04:42.0784864Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:42.0785237Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:04:42.0785605Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:04:42.0785752Z 2025-09-07T08:04:42.0785849Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0786340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0786714Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0787073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0787446Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0787819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 364, in forward 2025-09-07T08:04:42.0788219Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:04:42.0788620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 348, in forward 2025-09-07T08:04:42.0788980Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0789113Z 2025-09-07T08:04:42.0789210Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0789570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0789939Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0790295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0790676Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0791055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0791451Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0791846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0792217Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0792587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 271, in forward 2025-09-07T08:04:42.0792953Z mixed_query_layer = self.query(hidden_states) 2025-09-07T08:04:42.0793097Z 2025-09-07T08:04:42.0793195Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0793563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0793923Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0794280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0794658Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0795030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0795429Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0795824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0796273Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0796638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 272, in forward 2025-09-07T08:04:42.0796993Z mixed_key_layer = self.key(hidden_states) 2025-09-07T08:04:42.0797128Z 2025-09-07T08:04:42.0797220Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0797587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0797958Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0798311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0798708Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0799086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0799546Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0799951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0800322Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0800692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 273, in forward 2025-09-07T08:04:42.0801055Z mixed_value_layer = self.value(hidden_states) 2025-09-07T08:04:42.0801195Z 2025-09-07T08:04:42.0801276Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0801473Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0801688Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0802054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0802414Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0802789Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0803303Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0803682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0804080Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0804476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 321, in forward 2025-09-07T08:04:42.0804864Z attention_output = self.output(self_output, input_tensor) 2025-09-07T08:04:42.0805239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 307, in forward 2025-09-07T08:04:42.0805593Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0805730Z 2025-09-07T08:04:42.0805881Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0806254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0806614Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0806961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0807341Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0807713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0808107Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0808484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 335, in forward 2025-09-07T08:04:42.0808834Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0808970Z 2025-09-07T08:04:42.0809057Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0809538Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0809902Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0810253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0810638Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0811015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0811405Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0811787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:04:42.0812158Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:42.0812627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:04:42.0813003Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:04:42.0813153Z 2025-09-07T08:04:42.0813251Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0813622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0813982Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0814341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0814718Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0815101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 364, in forward 2025-09-07T08:04:42.0815507Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:04:42.0815902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 348, in forward 2025-09-07T08:04:42.0816266Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0816404Z 2025-09-07T08:04:42.0816502Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0816873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0817230Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0817646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0818033Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0818414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0818817Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0819212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0819586Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0819956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 271, in forward 2025-09-07T08:04:42.0820325Z mixed_query_layer = self.query(hidden_states) 2025-09-07T08:04:42.0820466Z 2025-09-07T08:04:42.0820566Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0820929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0821291Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0821652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0822035Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0822497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0822895Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0823292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0823665Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0824033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 272, in forward 2025-09-07T08:04:42.0824378Z mixed_key_layer = self.key(hidden_states) 2025-09-07T08:04:42.0824515Z 2025-09-07T08:04:42.0824606Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0824990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0825370Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0825824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0826221Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0826615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0827026Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0827425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0827822Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0828188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 273, in forward 2025-09-07T08:04:42.0828555Z mixed_value_layer = self.value(hidden_states) 2025-09-07T08:04:42.0828704Z 2025-09-07T08:04:42.0828784Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0828991Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0829212Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0829587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0829960Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0830339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0830740Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0831122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0831528Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0831930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 321, in forward 2025-09-07T08:04:42.0832328Z attention_output = self.output(self_output, input_tensor) 2025-09-07T08:04:42.0832720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 307, in forward 2025-09-07T08:04:42.0833075Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0833216Z 2025-09-07T08:04:42.0833314Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0833705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0834073Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0834432Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0834814Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0835199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0835600Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0836090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 335, in forward 2025-09-07T08:04:42.0836443Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0836588Z 2025-09-07T08:04:42.0836682Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0837051Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0837419Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0837782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0838163Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0838546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0838935Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0839383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:04:42.0839792Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:42.0840170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:04:42.0840545Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:04:42.0840702Z 2025-09-07T08:04:42.0840798Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0841173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0841539Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0841895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0842282Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0842671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 364, in forward 2025-09-07T08:04:42.0843077Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:04:42.0843471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 348, in forward 2025-09-07T08:04:42.0843829Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0843969Z 2025-09-07T08:04:42.0844059Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0844424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0844795Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0845147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0845534Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0845913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0846319Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0846717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0847087Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0847463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 271, in forward 2025-09-07T08:04:42.0847828Z mixed_query_layer = self.query(hidden_states) 2025-09-07T08:04:42.0847973Z 2025-09-07T08:04:42.0848069Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0848440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0848886Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0849245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0849630Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0850011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0850404Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0850798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0851167Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0851539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 272, in forward 2025-09-07T08:04:42.0851892Z mixed_key_layer = self.key(hidden_states) 2025-09-07T08:04:42.0852025Z 2025-09-07T08:04:42.0852178Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0852549Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0852909Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0853266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0853647Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0854020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0854413Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0854806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0855179Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0855556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 273, in forward 2025-09-07T08:04:42.0855909Z mixed_value_layer = self.value(hidden_states) 2025-09-07T08:04:42.0856052Z 2025-09-07T08:04:42.0856127Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0856327Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0856545Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0856908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0857269Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0857689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0858074Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0858455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0858851Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0859243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 321, in forward 2025-09-07T08:04:42.0859631Z attention_output = self.output(self_output, input_tensor) 2025-09-07T08:04:42.0860008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 307, in forward 2025-09-07T08:04:42.0860364Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0860495Z 2025-09-07T08:04:42.0860586Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0860954Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0861314Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0861673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0862157Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0862539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0862931Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0863318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 335, in forward 2025-09-07T08:04:42.0863672Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0863804Z 2025-09-07T08:04:42.0863896Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0864264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0864623Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0864981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0865423Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0865801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0866186Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0866576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:04:42.0866956Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:42.0867336Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:04:42.0867699Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:04:42.0867854Z 2025-09-07T08:04:42.0867945Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0868317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0868680Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0869032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0869413Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0869793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 364, in forward 2025-09-07T08:04:42.0870195Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:04:42.0870592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 348, in forward 2025-09-07T08:04:42.0870939Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0871078Z 2025-09-07T08:04:42.0871169Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0871537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0871924Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0872284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0872655Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0873037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0873438Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0873839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0874211Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0874575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 271, in forward 2025-09-07T08:04:42.0875017Z mixed_query_layer = self.query(hidden_states) 2025-09-07T08:04:42.0875163Z 2025-09-07T08:04:42.0875255Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0875620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0875981Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0876331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0876713Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0877090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0877485Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0877879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0878333Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0878717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 272, in forward 2025-09-07T08:04:42.0879078Z mixed_key_layer = self.key(hidden_states) 2025-09-07T08:04:42.0879213Z 2025-09-07T08:04:42.0879316Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0879688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0880061Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0880424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0880814Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0881206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0881610Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0882005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0882381Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0882751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 273, in forward 2025-09-07T08:04:42.0883112Z mixed_value_layer = self.value(hidden_states) 2025-09-07T08:04:42.0883255Z 2025-09-07T08:04:42.0883329Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0883531Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0883752Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0884120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0884481Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0884845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0885233Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0885614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0886011Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0886400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 321, in forward 2025-09-07T08:04:42.0886791Z attention_output = self.output(self_output, input_tensor) 2025-09-07T08:04:42.0887169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 307, in forward 2025-09-07T08:04:42.0887516Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0887645Z 2025-09-07T08:04:42.0887744Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0888187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0888553Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0888909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0889288Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0889657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0890040Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0890419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 335, in forward 2025-09-07T08:04:42.0890771Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0890901Z 2025-09-07T08:04:42.0890997Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0891425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0891794Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0892147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0892526Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0892906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0893283Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0893660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:04:42.0894037Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:42.0894404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:04:42.0894773Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:04:42.0894921Z 2025-09-07T08:04:42.0895013Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0895388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0895750Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0896104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0896476Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0896851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 364, in forward 2025-09-07T08:04:42.0905749Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:04:42.0906235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 348, in forward 2025-09-07T08:04:42.0906626Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0906768Z 2025-09-07T08:04:42.0906880Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0907274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0907655Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0908023Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0908418Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0908804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0909213Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0909618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0910180Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0910548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 271, in forward 2025-09-07T08:04:42.0910900Z mixed_query_layer = self.query(hidden_states) 2025-09-07T08:04:42.0911044Z 2025-09-07T08:04:42.0911144Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0911514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0911874Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0912226Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0912607Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0913082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0913489Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0913883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0914251Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0914607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 272, in forward 2025-09-07T08:04:42.0914957Z mixed_key_layer = self.key(hidden_states) 2025-09-07T08:04:42.0915091Z 2025-09-07T08:04:42.0915182Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0915548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0915905Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0916254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0916632Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0917007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0917397Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0917782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 320, in forward 2025-09-07T08:04:42.0918146Z self_output = self.self(input_tensor, attention_mask) 2025-09-07T08:04:42.0918504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 273, in forward 2025-09-07T08:04:42.0918864Z mixed_value_layer = self.value(hidden_states) 2025-09-07T08:04:42.0919004Z 2025-09-07T08:04:42.0919088Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0919286Z cudagraph partition due to non gpu ops 2025-09-07T08:04:42.0919505Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0919873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0920237Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0920584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0920961Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0921334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 362, in forward 2025-09-07T08:04:42.0921725Z attention_output = self.attention(hidden_states, attention_mask) 2025-09-07T08:04:42.0922115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 321, in forward 2025-09-07T08:04:42.0922498Z attention_output = self.output(self_output, input_tensor) 2025-09-07T08:04:42.0922974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 307, in forward 2025-09-07T08:04:42.0923345Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0923479Z 2025-09-07T08:04:42.0923576Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0923949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0924304Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0924661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0925036Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0925421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0925805Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0926245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 335, in forward 2025-09-07T08:04:42.0926601Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0926739Z 2025-09-07T08:04:42.0926831Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0927216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0927571Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0927921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0928295Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0928667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 363, in forward 2025-09-07T08:04:42.0929059Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:04:42.0929434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 336, in forward 2025-09-07T08:04:42.0929811Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:04:42.0930189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 150, in gelu 2025-09-07T08:04:42.0930551Z return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) 2025-09-07T08:04:42.0930704Z 2025-09-07T08:04:42.0930802Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0931173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 513, in forward 2025-09-07T08:04:42.0931542Z encoded_layers = self.encoder(embedding_output, 2025-09-07T08:04:42.0931900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 384, in forward 2025-09-07T08:04:42.0932287Z hidden_states = layer_module(hidden_states, attention_mask) 2025-09-07T08:04:42.0932665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 364, in forward 2025-09-07T08:04:42.0933064Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:04:42.0933461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 348, in forward 2025-09-07T08:04:42.0933815Z hidden_states = self.dense(hidden_states) 2025-09-07T08:04:42.0933945Z 2025-09-07T08:04:42.0934038Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0934410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 518, in forward 2025-09-07T08:04:42.0934773Z pooled_output = self.pooler(sequence_output) 2025-09-07T08:04:42.0935133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 404, in forward 2025-09-07T08:04:42.0935515Z pooled_output = self.dense(first_token_tensor) 2025-09-07T08:04:42.0935756Z 2025-09-07T08:04:42.0935854Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:42.0936225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 518, in forward 2025-09-07T08:04:42.0936587Z pooled_output = self.pooler(sequence_output) 2025-09-07T08:04:42.0936936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/modules/encoder/bert.py", line 405, in forward 2025-09-07T08:04:42.0937302Z pooled_output = self.activation(pooled_output) 2025-09-07T08:04:42.0937446Z 2025-09-07T08:04:44.1964142Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:44.1964722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 140, in torch_dynamo_resume_in_forward_at_137 2025-09-07T08:04:44.1965189Z return self.dropout(outputs) 2025-09-07T08:04:44.1965967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/embedding.py", line 160, in dropout 2025-09-07T08:04:44.1966358Z return self.dropout_layer(words) 2025-09-07T08:04:44.1966497Z 2025-09-07T08:04:44.2865852Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:44.2866329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/models/bert.py", line 266, in torch_dynamo_resume_in_forward_at_265 2025-09-07T08:04:44.2866816Z logits = self.qa_outputs(sequence_output) # [batch_size, seq_len, num_labels] 2025-09-07T08:04:44.2867025Z 2025-09-07T08:04:44.5107025Z pass 2025-09-07T08:04:44.5110279Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:46.7284807Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:04:46.7285716Z import pynvml # type: ignore[import] 2025-09-07T08:04:48.7399912Z 2025-09-07T08:04:49.1702545Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:04:49.1702847Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:04:49.1739782Z cpu eval functorch_dp_cifar10 2025-09-07T08:04:49.2401780Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:49.2675894Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:49.2900769Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:04:52.7216478Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7216787Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7217009Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7217216Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7217423Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7217728Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7217923Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7218104Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7218296Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7218504Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7218700Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7218896Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7219085Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7219282Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7219499Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7219692Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7219885Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7220066Z cudagraph partition due to non gpu ops 2025-09-07T08:04:52.7220301Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7220693Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7221480Z return mod(*inputs) 2025-09-07T08:04:52.7221836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7222200Z return self._forward_impl(x) 2025-09-07T08:04:52.7222556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 268, in _forward_impl 2025-09-07T08:04:52.7222912Z x = self.conv1(x) 2025-09-07T08:04:52.7223008Z 2025-09-07T08:04:52.7223121Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7223481Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7223807Z return mod(*inputs) 2025-09-07T08:04:52.7224121Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7224467Z return self._forward_impl(x) 2025-09-07T08:04:52.7224947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 270, in _forward_impl 2025-09-07T08:04:52.7225306Z x = self.relu(x) 2025-09-07T08:04:52.7225404Z 2025-09-07T08:04:52.7225497Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7225842Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7226161Z return mod(*inputs) 2025-09-07T08:04:52.7226462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7226800Z return self._forward_impl(x) 2025-09-07T08:04:52.7227135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 271, in _forward_impl 2025-09-07T08:04:52.7227484Z x = self.maxpool(x) 2025-09-07T08:04:52.7227572Z 2025-09-07T08:04:52.7227667Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7228013Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7228324Z return mod(*inputs) 2025-09-07T08:04:52.7228626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7228963Z return self._forward_impl(x) 2025-09-07T08:04:52.7229289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:04:52.7229636Z x = self.layer1(x) 2025-09-07T08:04:52.7229949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:04:52.7230279Z out = self.conv1(x) 2025-09-07T08:04:52.7230368Z 2025-09-07T08:04:52.7230461Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7230816Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7231127Z return mod(*inputs) 2025-09-07T08:04:52.7231441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7231779Z return self._forward_impl(x) 2025-09-07T08:04:52.7232111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:04:52.7232460Z x = self.layer1(x) 2025-09-07T08:04:52.7232759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 94, in forward 2025-09-07T08:04:52.7233097Z out = self.relu(out) 2025-09-07T08:04:52.7233200Z 2025-09-07T08:04:52.7233293Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7233638Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7233949Z return mod(*inputs) 2025-09-07T08:04:52.7234253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7234677Z return self._forward_impl(x) 2025-09-07T08:04:52.7235012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:04:52.7235355Z x = self.layer1(x) 2025-09-07T08:04:52.7235652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:04:52.7235985Z out = self.conv2(out) 2025-09-07T08:04:52.7236086Z 2025-09-07T08:04:52.7236185Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7236518Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7236827Z return mod(*inputs) 2025-09-07T08:04:52.7237128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7237463Z return self._forward_impl(x) 2025-09-07T08:04:52.7237852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:04:52.7238206Z x = self.layer1(x) 2025-09-07T08:04:52.7238504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 103, in forward 2025-09-07T08:04:52.7238837Z out = self.relu(out) 2025-09-07T08:04:52.7238937Z 2025-09-07T08:04:52.7239037Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7239366Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7239670Z return mod(*inputs) 2025-09-07T08:04:52.7239967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7240297Z return self._forward_impl(x) 2025-09-07T08:04:52.7240620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:04:52.7240965Z x = self.layer1(x) 2025-09-07T08:04:52.7241272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:04:52.7241597Z out = self.conv1(x) 2025-09-07T08:04:52.7241689Z 2025-09-07T08:04:52.7241785Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7242154Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7242464Z return mod(*inputs) 2025-09-07T08:04:52.7242765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7243102Z return self._forward_impl(x) 2025-09-07T08:04:52.7243453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:04:52.7243809Z x = self.layer1(x) 2025-09-07T08:04:52.7244123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 94, in forward 2025-09-07T08:04:52.7244463Z out = self.relu(out) 2025-09-07T08:04:52.7244567Z 2025-09-07T08:04:52.7244676Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7245032Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7245349Z return mod(*inputs) 2025-09-07T08:04:52.7245664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7246013Z return self._forward_impl(x) 2025-09-07T08:04:52.7246351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:04:52.7246702Z x = self.layer1(x) 2025-09-07T08:04:52.7247001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:04:52.7247333Z out = self.conv2(out) 2025-09-07T08:04:52.7247435Z 2025-09-07T08:04:52.7247534Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7247954Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7248266Z return mod(*inputs) 2025-09-07T08:04:52.7248574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7248916Z return self._forward_impl(x) 2025-09-07T08:04:52.7249247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 273, in _forward_impl 2025-09-07T08:04:52.7249588Z x = self.layer1(x) 2025-09-07T08:04:52.7249884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 103, in forward 2025-09-07T08:04:52.7250218Z out = self.relu(out) 2025-09-07T08:04:52.7250314Z 2025-09-07T08:04:52.7250412Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7250744Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7251126Z return mod(*inputs) 2025-09-07T08:04:52.7251430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7251762Z return self._forward_impl(x) 2025-09-07T08:04:52.7252089Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:04:52.7252435Z x = self.layer2(x) 2025-09-07T08:04:52.7252731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:04:52.7253060Z out = self.conv1(x) 2025-09-07T08:04:52.7253150Z 2025-09-07T08:04:52.7253248Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7253576Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7253882Z return mod(*inputs) 2025-09-07T08:04:52.7254178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7254516Z return self._forward_impl(x) 2025-09-07T08:04:52.7254838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:04:52.7255180Z x = self.layer2(x) 2025-09-07T08:04:52.7255470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 94, in forward 2025-09-07T08:04:52.7255798Z out = self.relu(out) 2025-09-07T08:04:52.7255891Z 2025-09-07T08:04:52.7255989Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7256313Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7256620Z return mod(*inputs) 2025-09-07T08:04:52.7256917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7257247Z return self._forward_impl(x) 2025-09-07T08:04:52.7257653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:04:52.7257998Z x = self.layer2(x) 2025-09-07T08:04:52.7258298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:04:52.7258650Z out = self.conv2(out) 2025-09-07T08:04:52.7258754Z 2025-09-07T08:04:52.7258855Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7259205Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7259521Z return mod(*inputs) 2025-09-07T08:04:52.7259837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7260176Z return self._forward_impl(x) 2025-09-07T08:04:52.7260509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:04:52.7260943Z x = self.layer2(x) 2025-09-07T08:04:52.7261244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 100, in forward 2025-09-07T08:04:52.7261589Z identity = self.downsample(x) 2025-09-07T08:04:52.7261707Z 2025-09-07T08:04:52.7261809Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7262145Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7262457Z return mod(*inputs) 2025-09-07T08:04:52.7262765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7263103Z return self._forward_impl(x) 2025-09-07T08:04:52.7263435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:04:52.7263783Z x = self.layer2(x) 2025-09-07T08:04:52.7264101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 103, in forward 2025-09-07T08:04:52.7264515Z out = self.relu(out) 2025-09-07T08:04:52.7264616Z 2025-09-07T08:04:52.7264724Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7265069Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7265392Z return mod(*inputs) 2025-09-07T08:04:52.7265707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7266053Z return self._forward_impl(x) 2025-09-07T08:04:52.7266393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:04:52.7266742Z x = self.layer2(x) 2025-09-07T08:04:52.7267050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:04:52.7267382Z out = self.conv1(x) 2025-09-07T08:04:52.7267475Z 2025-09-07T08:04:52.7267582Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7267924Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7268235Z return mod(*inputs) 2025-09-07T08:04:52.7268540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7268876Z return self._forward_impl(x) 2025-09-07T08:04:52.7269205Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:04:52.7269569Z x = self.layer2(x) 2025-09-07T08:04:52.7269881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 94, in forward 2025-09-07T08:04:52.7270214Z out = self.relu(out) 2025-09-07T08:04:52.7270313Z 2025-09-07T08:04:52.7270412Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7270750Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7271062Z return mod(*inputs) 2025-09-07T08:04:52.7271361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7271700Z return self._forward_impl(x) 2025-09-07T08:04:52.7272029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:04:52.7272369Z x = self.layer2(x) 2025-09-07T08:04:52.7272662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:04:52.7272992Z out = self.conv2(out) 2025-09-07T08:04:52.7273096Z 2025-09-07T08:04:52.7273194Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7273523Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7273830Z return mod(*inputs) 2025-09-07T08:04:52.7274232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7274566Z return self._forward_impl(x) 2025-09-07T08:04:52.7274889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 274, in _forward_impl 2025-09-07T08:04:52.7275243Z x = self.layer2(x) 2025-09-07T08:04:52.7275542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 103, in forward 2025-09-07T08:04:52.7275873Z out = self.relu(out) 2025-09-07T08:04:52.7275967Z 2025-09-07T08:04:52.7276066Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7276391Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7276700Z return mod(*inputs) 2025-09-07T08:04:52.7276995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7277335Z return self._forward_impl(x) 2025-09-07T08:04:52.7277725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:04:52.7278080Z x = self.layer3(x) 2025-09-07T08:04:52.7278383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:04:52.7278708Z out = self.conv1(x) 2025-09-07T08:04:52.7278798Z 2025-09-07T08:04:52.7278892Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7279219Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7279525Z return mod(*inputs) 2025-09-07T08:04:52.7279820Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7280154Z return self._forward_impl(x) 2025-09-07T08:04:52.7280484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:04:52.7280831Z x = self.layer3(x) 2025-09-07T08:04:52.7281119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 94, in forward 2025-09-07T08:04:52.7281445Z out = self.relu(out) 2025-09-07T08:04:52.7281536Z 2025-09-07T08:04:52.7281631Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7281960Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7282266Z return mod(*inputs) 2025-09-07T08:04:52.7282559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7282891Z return self._forward_impl(x) 2025-09-07T08:04:52.7283210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:04:52.7283553Z x = self.layer3(x) 2025-09-07T08:04:52.7283852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:04:52.7284191Z out = self.conv2(out) 2025-09-07T08:04:52.7284287Z 2025-09-07T08:04:52.7284385Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7284715Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7285027Z return mod(*inputs) 2025-09-07T08:04:52.7285326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7285658Z return self._forward_impl(x) 2025-09-07T08:04:52.7285981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:04:52.7286321Z x = self.layer3(x) 2025-09-07T08:04:52.7286615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 100, in forward 2025-09-07T08:04:52.7286956Z identity = self.downsample(x) 2025-09-07T08:04:52.7287145Z 2025-09-07T08:04:52.7287248Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7287578Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7287905Z return mod(*inputs) 2025-09-07T08:04:52.7288224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7288574Z return self._forward_impl(x) 2025-09-07T08:04:52.7288919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:04:52.7289277Z x = self.layer3(x) 2025-09-07T08:04:52.7289587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 103, in forward 2025-09-07T08:04:52.7289924Z out = self.relu(out) 2025-09-07T08:04:52.7290022Z 2025-09-07T08:04:52.7290129Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7290555Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7290879Z return mod(*inputs) 2025-09-07T08:04:52.7291185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7291526Z return self._forward_impl(x) 2025-09-07T08:04:52.7291857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:04:52.7292202Z x = self.layer3(x) 2025-09-07T08:04:52.7292501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:04:52.7292830Z out = self.conv1(x) 2025-09-07T08:04:52.7292922Z 2025-09-07T08:04:52.7293019Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7293355Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7293672Z return mod(*inputs) 2025-09-07T08:04:52.7293972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7294307Z return self._forward_impl(x) 2025-09-07T08:04:52.7294632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:04:52.7294977Z x = self.layer3(x) 2025-09-07T08:04:52.7295268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 94, in forward 2025-09-07T08:04:52.7295594Z out = self.relu(out) 2025-09-07T08:04:52.7295686Z 2025-09-07T08:04:52.7295783Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7296110Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7296416Z return mod(*inputs) 2025-09-07T08:04:52.7296715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7297052Z return self._forward_impl(x) 2025-09-07T08:04:52.7297375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:04:52.7297786Z x = self.layer3(x) 2025-09-07T08:04:52.7298094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:04:52.7298428Z out = self.conv2(out) 2025-09-07T08:04:52.7298532Z 2025-09-07T08:04:52.7298628Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7298964Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7299275Z return mod(*inputs) 2025-09-07T08:04:52.7299649Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7300010Z return self._forward_impl(x) 2025-09-07T08:04:52.7300369Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 275, in _forward_impl 2025-09-07T08:04:52.7300808Z x = self.layer3(x) 2025-09-07T08:04:52.7301120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 103, in forward 2025-09-07T08:04:52.7301461Z out = self.relu(out) 2025-09-07T08:04:52.7301564Z 2025-09-07T08:04:52.7301673Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7302034Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7302357Z return mod(*inputs) 2025-09-07T08:04:52.7302663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7303004Z return self._forward_impl(x) 2025-09-07T08:04:52.7303538Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:04:52.7303889Z x = self.layer4(x) 2025-09-07T08:04:52.7304298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:04:52.7304642Z out = self.conv1(x) 2025-09-07T08:04:52.7304736Z 2025-09-07T08:04:52.7304839Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7305180Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7305498Z return mod(*inputs) 2025-09-07T08:04:52.7305796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7306137Z return self._forward_impl(x) 2025-09-07T08:04:52.7306481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:04:52.7306850Z x = self.layer4(x) 2025-09-07T08:04:52.7307147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 94, in forward 2025-09-07T08:04:52.7307487Z out = self.relu(out) 2025-09-07T08:04:52.7307582Z 2025-09-07T08:04:52.7307684Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7308012Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7308321Z return mod(*inputs) 2025-09-07T08:04:52.7308622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7308957Z return self._forward_impl(x) 2025-09-07T08:04:52.7309285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:04:52.7309629Z x = self.layer4(x) 2025-09-07T08:04:52.7309919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:04:52.7310250Z out = self.conv2(out) 2025-09-07T08:04:52.7310352Z 2025-09-07T08:04:52.7310447Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7310782Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7311088Z return mod(*inputs) 2025-09-07T08:04:52.7311380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7311715Z return self._forward_impl(x) 2025-09-07T08:04:52.7312042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:04:52.7312385Z x = self.layer4(x) 2025-09-07T08:04:52.7312682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 100, in forward 2025-09-07T08:04:52.7313024Z identity = self.downsample(x) 2025-09-07T08:04:52.7313142Z 2025-09-07T08:04:52.7313239Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7313568Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7313997Z return mod(*inputs) 2025-09-07T08:04:52.7314299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7314638Z return self._forward_impl(x) 2025-09-07T08:04:52.7314971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:04:52.7315316Z x = self.layer4(x) 2025-09-07T08:04:52.7315614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 103, in forward 2025-09-07T08:04:52.7315946Z out = self.relu(out) 2025-09-07T08:04:52.7316044Z 2025-09-07T08:04:52.7316143Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7316471Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7316777Z return mod(*inputs) 2025-09-07T08:04:52.7317153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7317504Z return self._forward_impl(x) 2025-09-07T08:04:52.7317832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:04:52.7318178Z x = self.layer4(x) 2025-09-07T08:04:52.7318473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 92, in forward 2025-09-07T08:04:52.7318802Z out = self.conv1(x) 2025-09-07T08:04:52.7318895Z 2025-09-07T08:04:52.7318994Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7319322Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7319628Z return mod(*inputs) 2025-09-07T08:04:52.7319930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7320268Z return self._forward_impl(x) 2025-09-07T08:04:52.7320614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:04:52.7320955Z x = self.layer4(x) 2025-09-07T08:04:52.7321251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 94, in forward 2025-09-07T08:04:52.7321582Z out = self.relu(out) 2025-09-07T08:04:52.7321674Z 2025-09-07T08:04:52.7321772Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7322100Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7322407Z return mod(*inputs) 2025-09-07T08:04:52.7322703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7323033Z return self._forward_impl(x) 2025-09-07T08:04:52.7323356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 276, in _forward_impl 2025-09-07T08:04:52.7323703Z x = self.layer4(x) 2025-09-07T08:04:52.7323993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 96, in forward 2025-09-07T08:04:52.7324321Z out = self.conv2(out) 2025-09-07T08:04:52.7324418Z 2025-09-07T08:04:52.7324515Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7324844Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7325148Z return mod(*inputs) 2025-09-07T08:04:52.7325441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7325773Z return self._forward_impl(x) 2025-09-07T08:04:52.7326096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 278, in _forward_impl 2025-09-07T08:04:52.7326438Z x = self.avgpool(x) 2025-09-07T08:04:52.7326535Z 2025-09-07T08:04:52.7326624Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:04:52.7328960Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:04:52.7329273Z return mod(*inputs) 2025-09-07T08:04:52.7329569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 285, in forward 2025-09-07T08:04:52.7329907Z return self._forward_impl(x) 2025-09-07T08:04:52.7330239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 280, in _forward_impl 2025-09-07T08:04:52.7330588Z x = self.fc(x) 2025-09-07T08:04:52.7330675Z 2025-09-07T08:05:07.3918865Z pass 2025-09-07T08:05:07.3921368Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:05:09.3984271Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:05:09.3985165Z import pynvml # type: ignore[import] 2025-09-07T08:05:11.4081639Z 2025-09-07T08:05:11.5978676Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:05:11.5978967Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:05:11.5982875Z cpu eval functorch_maml_omniglot 2025-09-07T08:05:11.6079147Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:05:11.6157682Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:05:11.6205173Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:05:13.8168273Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:13.8168723Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:05:13.8169112Z return mod(*inputs) 2025-09-07T08:05:13.8169221Z 2025-09-07T08:05:13.8169313Z cudagraph partition due to non gpu ops 2025-09-07T08:05:13.8169531Z cudagraph partition due to non gpu ops 2025-09-07T08:05:13.8169759Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:13.8170124Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:05:13.8170456Z return mod(*inputs) 2025-09-07T08:05:13.8170551Z 2025-09-07T08:05:13.8170645Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:13.8171002Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:05:13.8171320Z return mod(*inputs) 2025-09-07T08:05:13.8171416Z 2025-09-07T08:05:13.8171520Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:13.8171866Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:05:13.8172182Z return mod(*inputs) 2025-09-07T08:05:13.8172290Z 2025-09-07T08:05:13.8172389Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:13.8172737Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:05:13.8173048Z return mod(*inputs) 2025-09-07T08:05:13.8173139Z 2025-09-07T08:05:13.8173231Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:13.8173579Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:05:13.8173896Z return mod(*inputs) 2025-09-07T08:05:13.8173991Z 2025-09-07T08:05:13.8174109Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:13.8174450Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:05:13.8174755Z return mod(*inputs) 2025-09-07T08:05:13.8174864Z 2025-09-07T08:05:13.8174961Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:13.8175701Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:05:13.8176020Z return mod(*inputs) 2025-09-07T08:05:13.8176116Z 2025-09-07T08:05:13.8176215Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:13.8176560Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:05:13.8176874Z return mod(*inputs) 2025-09-07T08:05:13.8176965Z 2025-09-07T08:05:13.8177066Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:13.8177403Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:05:13.8177772Z return mod(*inputs) 2025-09-07T08:05:13.8177868Z 2025-09-07T08:05:19.9970704Z pass 2025-09-07T08:05:19.9973758Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:05:21.9517851Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:05:21.9518862Z import pynvml # type: ignore[import] 2025-09-07T08:05:23.9561892Z 2025-09-07T08:05:25.6349066Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:05:25.6349351Z loading model: 0it [00:01, ?it/s] 2025-09-07T08:05:25.6400656Z cpu eval hf_Albert 2025-09-07T08:05:26.5545186Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:05:26.7960599Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:05:27.0941767Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:05:34.1765325Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1765801Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1766176Z return mod(**inputs) 2025-09-07T08:05:34.1766600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1767023Z outputs = self.albert( 2025-09-07T08:05:34.1767414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 700, in forward 2025-09-07T08:05:34.1767819Z embedding_output = self.embeddings( 2025-09-07T08:05:34.1768220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 240, in forward 2025-09-07T08:05:34.1768632Z embeddings = self.LayerNorm(embeddings) 2025-09-07T08:05:34.1768770Z 2025-09-07T08:05:34.1768855Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.1769051Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.1769247Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.1769460Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.1769650Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.1769833Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.1770022Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.1770214Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.1770431Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.1770620Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.1770814Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.1770998Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.1771230Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1771596Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1771925Z return mod(**inputs) 2025-09-07T08:05:34.1772294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1773084Z outputs = self.albert( 2025-09-07T08:05:34.1773455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1773851Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1774254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 520, in forward 2025-09-07T08:05:34.1774696Z hidden_states = self.embedding_hidden_mapping_in(hidden_states) 2025-09-07T08:05:34.1774887Z 2025-09-07T08:05:34.1774988Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1775348Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1775676Z return mod(**inputs) 2025-09-07T08:05:34.1776040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1776431Z outputs = self.albert( 2025-09-07T08:05:34.1776951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1777353Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1777796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1778240Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1778676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1779204Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1779729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1780331Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1780851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 380, in forward 2025-09-07T08:05:34.1781240Z self.query(hidden_states) 2025-09-07T08:05:34.1781366Z 2025-09-07T08:05:34.1781466Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1781826Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1782143Z return mod(**inputs) 2025-09-07T08:05:34.1782499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1782895Z outputs = self.albert( 2025-09-07T08:05:34.1783261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1783683Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1784073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1784493Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1784931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1785439Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1785992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1786498Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1786993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 385, in forward 2025-09-07T08:05:34.1787373Z self.key(hidden_states) 2025-09-07T08:05:34.1787577Z 2025-09-07T08:05:34.1787678Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1788070Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1788382Z return mod(**inputs) 2025-09-07T08:05:34.1788732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1789131Z outputs = self.albert( 2025-09-07T08:05:34.1789495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1789898Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1790276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1790714Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1791264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1791804Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1792334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1792843Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1793353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 390, in forward 2025-09-07T08:05:34.1793757Z self.value(hidden_states) 2025-09-07T08:05:34.1793879Z 2025-09-07T08:05:34.1793954Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.1794190Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1794546Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1794867Z return mod(**inputs) 2025-09-07T08:05:34.1795233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1795623Z outputs = self.albert( 2025-09-07T08:05:34.1795985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1796367Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1796746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1797176Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1797604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1798120Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1798633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1799137Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1799640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:05:34.1800100Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:34.1800294Z 2025-09-07T08:05:34.1800401Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1800743Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1801079Z return mod(**inputs) 2025-09-07T08:05:34.1801432Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1801902Z outputs = self.albert( 2025-09-07T08:05:34.1802286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1802680Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1803220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1803669Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1804101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1804617Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1805127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1805735Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1806256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 415, in forward 2025-09-07T08:05:34.1806680Z projected_context_layer = self.dense(attention_output) 2025-09-07T08:05:34.1806845Z 2025-09-07T08:05:34.1806953Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1807303Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1807624Z return mod(**inputs) 2025-09-07T08:05:34.1807997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1808390Z outputs = self.albert( 2025-09-07T08:05:34.1808748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1809139Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1809525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1809955Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1810378Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1810885Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1811399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.1811799Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.1812206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.1812606Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.1813007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 462, in ff_chunk 2025-09-07T08:05:34.1813413Z ffn_output = self.ffn(attention_output) 2025-09-07T08:05:34.1824776Z 2025-09-07T08:05:34.1824929Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1825341Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1825691Z return mod(**inputs) 2025-09-07T08:05:34.1826108Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1826524Z outputs = self.albert( 2025-09-07T08:05:34.1826905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1827303Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1827740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1828343Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1828788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1829315Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1829840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.1830245Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.1830648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.1831072Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.1831553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:05:34.1831986Z ffn_output = self.activation(ffn_output) 2025-09-07T08:05:34.1832348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:34.1832805Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:34.1833052Z 2025-09-07T08:05:34.1833157Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1833529Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1833856Z return mod(**inputs) 2025-09-07T08:05:34.1834231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1834620Z outputs = self.albert( 2025-09-07T08:05:34.1834987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1835385Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1835770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1836207Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1836623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1837147Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1837665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.1838073Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.1838470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.1838880Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.1839273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 464, in ff_chunk 2025-09-07T08:05:34.1839679Z ffn_output = self.ffn_output(ffn_output) 2025-09-07T08:05:34.1839813Z 2025-09-07T08:05:34.1839921Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1840270Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1840591Z return mod(**inputs) 2025-09-07T08:05:34.1840953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1841338Z outputs = self.albert( 2025-09-07T08:05:34.1841702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1842082Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1842545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1842971Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1843391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1843905Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1844409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1844917Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1845430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 380, in forward 2025-09-07T08:05:34.1845825Z self.query(hidden_states) 2025-09-07T08:05:34.1846007Z 2025-09-07T08:05:34.1846125Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1846475Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1846792Z return mod(**inputs) 2025-09-07T08:05:34.1847154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1847542Z outputs = self.albert( 2025-09-07T08:05:34.1847904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1848296Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1848674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1849096Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1849527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1850035Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1850553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1851064Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1851571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 385, in forward 2025-09-07T08:05:34.1851956Z self.key(hidden_states) 2025-09-07T08:05:34.1852064Z 2025-09-07T08:05:34.1852161Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1852508Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1852827Z return mod(**inputs) 2025-09-07T08:05:34.1853195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1853584Z outputs = self.albert( 2025-09-07T08:05:34.1853944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1854376Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1854742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1855168Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1855585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1856099Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1856710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1857218Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1857779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 390, in forward 2025-09-07T08:05:34.1858180Z self.value(hidden_states) 2025-09-07T08:05:34.1858305Z 2025-09-07T08:05:34.1858384Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.1858622Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1858977Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1859302Z return mod(**inputs) 2025-09-07T08:05:34.1859668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1860064Z outputs = self.albert( 2025-09-07T08:05:34.1860525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1860915Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1861301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1861728Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1862153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1862667Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1863178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1863690Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1864349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:05:34.1864814Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:34.1865009Z 2025-09-07T08:05:34.1865115Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1865467Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1865783Z return mod(**inputs) 2025-09-07T08:05:34.1866138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1866519Z outputs = self.albert( 2025-09-07T08:05:34.1866882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1867271Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1867652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1868087Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1868511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1869014Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1869526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1870028Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1870526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 415, in forward 2025-09-07T08:05:34.1871032Z projected_context_layer = self.dense(attention_output) 2025-09-07T08:05:34.1871196Z 2025-09-07T08:05:34.1871300Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1871641Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1871954Z return mod(**inputs) 2025-09-07T08:05:34.1872309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1872693Z outputs = self.albert( 2025-09-07T08:05:34.1873044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1873434Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1873812Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1874232Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1874724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1875232Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1875747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.1876142Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.1876543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.1876938Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.1877326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 462, in ff_chunk 2025-09-07T08:05:34.1877724Z ffn_output = self.ffn(attention_output) 2025-09-07T08:05:34.1877861Z 2025-09-07T08:05:34.1877962Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1878314Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1878623Z return mod(**inputs) 2025-09-07T08:05:34.1878978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1879363Z outputs = self.albert( 2025-09-07T08:05:34.1879720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1880108Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1880478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1880899Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1881321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1881833Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1882342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.1882734Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.1883124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.1883521Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.1883910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:05:34.1884348Z ffn_output = self.activation(ffn_output) 2025-09-07T08:05:34.1884699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:34.1886170Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:34.1886414Z 2025-09-07T08:05:34.1886510Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1886861Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1887176Z return mod(**inputs) 2025-09-07T08:05:34.1887535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1887916Z outputs = self.albert( 2025-09-07T08:05:34.1888278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1888666Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1889043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1889546Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1889968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1890480Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1890989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.1891382Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.1891778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.1892164Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.1892553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 464, in ff_chunk 2025-09-07T08:05:34.1892959Z ffn_output = self.ffn_output(ffn_output) 2025-09-07T08:05:34.1893091Z 2025-09-07T08:05:34.1893190Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1893537Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1893842Z return mod(**inputs) 2025-09-07T08:05:34.1894195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1894573Z outputs = self.albert( 2025-09-07T08:05:34.1894931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1895312Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1895684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1896105Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1896533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1897041Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1897615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1898124Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1898630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 380, in forward 2025-09-07T08:05:34.1899018Z self.query(hidden_states) 2025-09-07T08:05:34.1899124Z 2025-09-07T08:05:34.1899225Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1899564Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1899959Z return mod(**inputs) 2025-09-07T08:05:34.1900317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1900704Z outputs = self.albert( 2025-09-07T08:05:34.1901066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1901449Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1901829Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1902246Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1902660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1903329Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1903961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1904465Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1904972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 385, in forward 2025-09-07T08:05:34.1905360Z self.key(hidden_states) 2025-09-07T08:05:34.1905468Z 2025-09-07T08:05:34.1905564Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1905912Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1906225Z return mod(**inputs) 2025-09-07T08:05:34.1906585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1906971Z outputs = self.albert( 2025-09-07T08:05:34.1907322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1907706Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1908079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1908495Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1908912Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1909414Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1909918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1910418Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1910925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 390, in forward 2025-09-07T08:05:34.1911309Z self.value(hidden_states) 2025-09-07T08:05:34.1911414Z 2025-09-07T08:05:34.1911489Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.1911713Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1912051Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1912350Z return mod(**inputs) 2025-09-07T08:05:34.1912689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1913059Z outputs = self.albert( 2025-09-07T08:05:34.1913406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1913919Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1914295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1914710Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1915124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1915630Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1916133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1916624Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1917114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:05:34.1917643Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:34.1917842Z 2025-09-07T08:05:34.1917939Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1918287Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1918590Z return mod(**inputs) 2025-09-07T08:05:34.1918936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1919311Z outputs = self.albert( 2025-09-07T08:05:34.1919661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1920045Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1920412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1920827Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1921243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1921743Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1922241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1922743Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1923235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 415, in forward 2025-09-07T08:05:34.1923653Z projected_context_layer = self.dense(attention_output) 2025-09-07T08:05:34.1923845Z 2025-09-07T08:05:34.1923953Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1924305Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1924614Z return mod(**inputs) 2025-09-07T08:05:34.1924966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1925344Z outputs = self.albert( 2025-09-07T08:05:34.1925691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1926068Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1926431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1926842Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1927248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1927753Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1928336Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.1928728Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.1929118Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.1929514Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.1929895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 462, in ff_chunk 2025-09-07T08:05:34.1930285Z ffn_output = self.ffn(attention_output) 2025-09-07T08:05:34.1930411Z 2025-09-07T08:05:34.1930503Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1930837Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1931141Z return mod(**inputs) 2025-09-07T08:05:34.1931555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1931930Z outputs = self.albert( 2025-09-07T08:05:34.1932280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1932660Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1933033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1933443Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1933854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1934361Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1934871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.1935258Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.1935666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.1936049Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.1936427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:05:34.1936825Z ffn_output = self.activation(ffn_output) 2025-09-07T08:05:34.1937167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:34.1937667Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:34.1937901Z 2025-09-07T08:05:34.1938004Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1938351Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1938740Z return mod(**inputs) 2025-09-07T08:05:34.1939103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1939480Z outputs = self.albert( 2025-09-07T08:05:34.1939828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1940202Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1940565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1940978Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1941406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1942045Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1942565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.1942952Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.1943350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.1943753Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.1944136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 464, in ff_chunk 2025-09-07T08:05:34.1944549Z ffn_output = self.ffn_output(ffn_output) 2025-09-07T08:05:34.1944686Z 2025-09-07T08:05:34.1944786Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1945198Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1945531Z return mod(**inputs) 2025-09-07T08:05:34.1945891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1946292Z outputs = self.albert( 2025-09-07T08:05:34.1946686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1947089Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1947479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1947899Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1948314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1948842Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1949363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1949868Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1950365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 380, in forward 2025-09-07T08:05:34.1950737Z self.query(hidden_states) 2025-09-07T08:05:34.1950858Z 2025-09-07T08:05:34.1950955Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1951291Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1951608Z return mod(**inputs) 2025-09-07T08:05:34.1951952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1952351Z outputs = self.albert( 2025-09-07T08:05:34.1952706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1953105Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1953472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1953901Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1954311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1954820Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1955316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1955897Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1956388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 385, in forward 2025-09-07T08:05:34.1956775Z self.key(hidden_states) 2025-09-07T08:05:34.1956879Z 2025-09-07T08:05:34.1956985Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1957316Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1957618Z return mod(**inputs) 2025-09-07T08:05:34.1957970Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1958351Z outputs = self.albert( 2025-09-07T08:05:34.1958699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1959085Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1959532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1959979Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1960406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1960914Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1961429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1961928Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1962427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 390, in forward 2025-09-07T08:05:34.1962824Z self.value(hidden_states) 2025-09-07T08:05:34.1962934Z 2025-09-07T08:05:34.1963012Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.1963242Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1963590Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1963908Z return mod(**inputs) 2025-09-07T08:05:34.1964257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1964639Z outputs = self.albert( 2025-09-07T08:05:34.1965000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1965386Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1965759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1966170Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1966594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1967105Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1967615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1968118Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1968611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:05:34.1969069Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:34.1969271Z 2025-09-07T08:05:34.1969364Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1969815Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1970132Z return mod(**inputs) 2025-09-07T08:05:34.1970477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1970859Z outputs = self.albert( 2025-09-07T08:05:34.1971220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1971605Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1971969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1972389Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1972806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1973387Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1973899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.1974400Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.1974898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 415, in forward 2025-09-07T08:05:34.1975321Z projected_context_layer = self.dense(attention_output) 2025-09-07T08:05:34.1975490Z 2025-09-07T08:05:34.1975585Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1975933Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1976241Z return mod(**inputs) 2025-09-07T08:05:34.1976599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1976987Z outputs = self.albert( 2025-09-07T08:05:34.1977349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1977782Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1978156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1978580Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1979001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1979513Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1980021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.1980421Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.1980825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.1981222Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.1981612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 462, in ff_chunk 2025-09-07T08:05:34.1982011Z ffn_output = self.ffn(attention_output) 2025-09-07T08:05:34.1982138Z 2025-09-07T08:05:34.1982237Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1982587Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1982901Z return mod(**inputs) 2025-09-07T08:05:34.1983260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1983718Z outputs = self.albert( 2025-09-07T08:05:34.1984081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1984470Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1984843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1985261Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1985675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1986179Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1986689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.1987083Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.1987552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.1987943Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.1988330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:05:34.1988736Z ffn_output = self.activation(ffn_output) 2025-09-07T08:05:34.1989092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:34.1989547Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:34.1989782Z 2025-09-07T08:05:34.1989878Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1990231Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1990545Z return mod(**inputs) 2025-09-07T08:05:34.1990908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1991294Z outputs = self.albert( 2025-09-07T08:05:34.1991654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1992039Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1992412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1992833Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1993243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.1993751Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.1994262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.1994660Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.1995053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.1995438Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.1995819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 464, in ff_chunk 2025-09-07T08:05:34.1996216Z ffn_output = self.ffn_output(ffn_output) 2025-09-07T08:05:34.1996346Z 2025-09-07T08:05:34.1996445Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.1996789Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.1997098Z return mod(**inputs) 2025-09-07T08:05:34.1997452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.1997931Z outputs = self.albert( 2025-09-07T08:05:34.1998297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.1998674Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.1999050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.1999473Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.1999892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2000403Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2000903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2001486Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2001992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 380, in forward 2025-09-07T08:05:34.2002377Z self.query(hidden_states) 2025-09-07T08:05:34.2002484Z 2025-09-07T08:05:34.2002587Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2002924Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2003361Z return mod(**inputs) 2025-09-07T08:05:34.2003718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2004097Z outputs = self.albert( 2025-09-07T08:05:34.2004456Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2004845Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2005221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2005639Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2006060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2006564Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2007064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2007564Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2008072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 385, in forward 2025-09-07T08:05:34.2008461Z self.key(hidden_states) 2025-09-07T08:05:34.2008568Z 2025-09-07T08:05:34.2008668Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2009003Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2009312Z return mod(**inputs) 2025-09-07T08:05:34.2009668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2010049Z outputs = self.albert( 2025-09-07T08:05:34.2010404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2010789Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2011161Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2011584Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2012134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2012632Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2013138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2013636Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2014135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 390, in forward 2025-09-07T08:05:34.2014517Z self.value(hidden_states) 2025-09-07T08:05:34.2014621Z 2025-09-07T08:05:34.2014699Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.2014929Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2015400Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2015727Z return mod(**inputs) 2025-09-07T08:05:34.2016080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2016468Z outputs = self.albert( 2025-09-07T08:05:34.2016830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2017221Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2017653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2018078Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2018506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2019033Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2019548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2020050Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2020543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:05:34.2020998Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:34.2021200Z 2025-09-07T08:05:34.2021298Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2021650Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2021966Z return mod(**inputs) 2025-09-07T08:05:34.2022319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2022705Z outputs = self.albert( 2025-09-07T08:05:34.2023065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2023449Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2023819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2024239Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2024662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2025170Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2025680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2026272Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2026770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 415, in forward 2025-09-07T08:05:34.2027198Z projected_context_layer = self.dense(attention_output) 2025-09-07T08:05:34.2027366Z 2025-09-07T08:05:34.2027460Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2027805Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2028112Z return mod(**inputs) 2025-09-07T08:05:34.2028466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2028844Z outputs = self.albert( 2025-09-07T08:05:34.2029205Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2029713Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2030081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2030503Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2030921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2031429Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2031945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2032332Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2032728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2033128Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2033519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 462, in ff_chunk 2025-09-07T08:05:34.2033917Z ffn_output = self.ffn(attention_output) 2025-09-07T08:05:34.2034045Z 2025-09-07T08:05:34.2034139Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2034488Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2034805Z return mod(**inputs) 2025-09-07T08:05:34.2035169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2035555Z outputs = self.albert( 2025-09-07T08:05:34.2035927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2036316Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2036697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2037121Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2037533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2038044Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2038557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2038952Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2039346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2039732Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2040234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:05:34.2040641Z ffn_output = self.activation(ffn_output) 2025-09-07T08:05:34.2040993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:34.2041446Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:34.2041681Z 2025-09-07T08:05:34.2041780Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2042137Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2042459Z return mod(**inputs) 2025-09-07T08:05:34.2042816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2043201Z outputs = self.albert( 2025-09-07T08:05:34.2043650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2044049Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2044429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2044853Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2045265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2045786Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2046295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2046695Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2047100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2047487Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2047878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 464, in ff_chunk 2025-09-07T08:05:34.2048279Z ffn_output = self.ffn_output(ffn_output) 2025-09-07T08:05:34.2048411Z 2025-09-07T08:05:34.2048513Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2048867Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2049177Z return mod(**inputs) 2025-09-07T08:05:34.2049535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2049921Z outputs = self.albert( 2025-09-07T08:05:34.2050282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2050667Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2051046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2051467Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2051889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2052397Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2052909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2053426Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2053943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 380, in forward 2025-09-07T08:05:34.2054422Z self.query(hidden_states) 2025-09-07T08:05:34.2054538Z 2025-09-07T08:05:34.2054643Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2054990Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2055306Z return mod(**inputs) 2025-09-07T08:05:34.2055668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2056054Z outputs = self.albert( 2025-09-07T08:05:34.2056414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2056801Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2057179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2057975Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2058411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2058930Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2059441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2059946Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2060451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 385, in forward 2025-09-07T08:05:34.2060837Z self.key(hidden_states) 2025-09-07T08:05:34.2060945Z 2025-09-07T08:05:34.2061050Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2061399Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2061718Z return mod(**inputs) 2025-09-07T08:05:34.2062079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2062467Z outputs = self.albert( 2025-09-07T08:05:34.2062826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2063216Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2063595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2064023Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2064448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2064965Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2065483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2065985Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2066498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 390, in forward 2025-09-07T08:05:34.2066884Z self.value(hidden_states) 2025-09-07T08:05:34.2066991Z 2025-09-07T08:05:34.2067068Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.2067302Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2067650Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2067961Z return mod(**inputs) 2025-09-07T08:05:34.2068311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2068780Z outputs = self.albert( 2025-09-07T08:05:34.2069138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2069527Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2069904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2070319Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2070738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2071245Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2071757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2072357Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2072856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:05:34.2073315Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:34.2073519Z 2025-09-07T08:05:34.2073615Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2073969Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2074285Z return mod(**inputs) 2025-09-07T08:05:34.2074634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2075019Z outputs = self.albert( 2025-09-07T08:05:34.2075387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2075784Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2076160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2076580Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2077003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2077513Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2078023Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2078529Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2079029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 415, in forward 2025-09-07T08:05:34.2079458Z projected_context_layer = self.dense(attention_output) 2025-09-07T08:05:34.2079628Z 2025-09-07T08:05:34.2079731Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2080088Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2080405Z return mod(**inputs) 2025-09-07T08:05:34.2080755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2081133Z outputs = self.albert( 2025-09-07T08:05:34.2081491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2081877Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2082243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2082754Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2083180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2083697Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2084204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2084597Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2085007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2085403Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2085808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 462, in ff_chunk 2025-09-07T08:05:34.2086294Z ffn_output = self.ffn(attention_output) 2025-09-07T08:05:34.2086430Z 2025-09-07T08:05:34.2086535Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2086896Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2087222Z return mod(**inputs) 2025-09-07T08:05:34.2087588Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2087974Z outputs = self.albert( 2025-09-07T08:05:34.2088343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2088734Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2089119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2089548Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2089974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2090496Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2091020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2091423Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2091824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2092215Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2092606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:05:34.2093013Z ffn_output = self.activation(ffn_output) 2025-09-07T08:05:34.2093370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:34.2093829Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:34.2094061Z 2025-09-07T08:05:34.2094164Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2094520Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2094839Z return mod(**inputs) 2025-09-07T08:05:34.2095202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2095587Z outputs = self.albert( 2025-09-07T08:05:34.2095940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2096329Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2096717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2097232Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2097684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2098198Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2098713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2099110Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2099505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2099892Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2100374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 464, in ff_chunk 2025-09-07T08:05:34.2100790Z ffn_output = self.ffn_output(ffn_output) 2025-09-07T08:05:34.2100922Z 2025-09-07T08:05:34.2101025Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2101372Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2101681Z return mod(**inputs) 2025-09-07T08:05:34.2102038Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2102419Z outputs = self.albert( 2025-09-07T08:05:34.2102779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2103284Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2103658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2104097Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2104518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2105028Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2105530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2106048Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2106545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 380, in forward 2025-09-07T08:05:34.2106932Z self.query(hidden_states) 2025-09-07T08:05:34.2107042Z 2025-09-07T08:05:34.2107143Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2107485Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2107792Z return mod(**inputs) 2025-09-07T08:05:34.2108142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2108517Z outputs = self.albert( 2025-09-07T08:05:34.2108872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2109252Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2109621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2110036Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2110451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2111151Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2111653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2112151Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2112647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 385, in forward 2025-09-07T08:05:34.2113028Z self.key(hidden_states) 2025-09-07T08:05:34.2113136Z 2025-09-07T08:05:34.2113235Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2113572Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2113881Z return mod(**inputs) 2025-09-07T08:05:34.2114229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2114709Z outputs = self.albert( 2025-09-07T08:05:34.2115058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2115443Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2115813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2116232Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2116646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2117149Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2117647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2118146Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2118649Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 390, in forward 2025-09-07T08:05:34.2119036Z self.value(hidden_states) 2025-09-07T08:05:34.2119140Z 2025-09-07T08:05:34.2119222Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.2119808Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2120264Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2129756Z return mod(**inputs) 2025-09-07T08:05:34.2130208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2130639Z outputs = self.albert( 2025-09-07T08:05:34.2131038Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2131448Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2131849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2132284Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2132722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2133248Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2133769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2134276Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2134793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:05:34.2135402Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:34.2135605Z 2025-09-07T08:05:34.2135718Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2136086Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2136410Z return mod(**inputs) 2025-09-07T08:05:34.2136769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2137159Z outputs = self.albert( 2025-09-07T08:05:34.2137598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2137986Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2138363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2138890Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2139320Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2139838Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2140355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2140863Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2141363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 415, in forward 2025-09-07T08:05:34.2141790Z projected_context_layer = self.dense(attention_output) 2025-09-07T08:05:34.2141965Z 2025-09-07T08:05:34.2142066Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2142432Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2142746Z return mod(**inputs) 2025-09-07T08:05:34.2143107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2143495Z outputs = self.albert( 2025-09-07T08:05:34.2143855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2144247Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2144615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2145038Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2145460Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2145980Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2146488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2146890Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2147298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2147698Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2148093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 462, in ff_chunk 2025-09-07T08:05:34.2148497Z ffn_output = self.ffn(attention_output) 2025-09-07T08:05:34.2148625Z 2025-09-07T08:05:34.2148726Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2149085Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2149508Z return mod(**inputs) 2025-09-07T08:05:34.2149869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2150250Z outputs = self.albert( 2025-09-07T08:05:34.2150615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2151005Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2151394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2151823Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2152243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2152845Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2153367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2153763Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2154163Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2154541Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2154916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:05:34.2155319Z ffn_output = self.activation(ffn_output) 2025-09-07T08:05:34.2155659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:34.2156110Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:34.2156344Z 2025-09-07T08:05:34.2156443Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2156787Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2157089Z return mod(**inputs) 2025-09-07T08:05:34.2157438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2157812Z outputs = self.albert( 2025-09-07T08:05:34.2158160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2158539Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2158907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2159318Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2159734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2160234Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2160736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2161118Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2161498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2161881Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2162256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 464, in ff_chunk 2025-09-07T08:05:34.2162649Z ffn_output = self.ffn_output(ffn_output) 2025-09-07T08:05:34.2162778Z 2025-09-07T08:05:34.2162875Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2163303Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2163609Z return mod(**inputs) 2025-09-07T08:05:34.2163955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2164332Z outputs = self.albert( 2025-09-07T08:05:34.2164685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2165068Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2165441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2165857Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2166269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2166839Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2167345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2167852Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2168349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 380, in forward 2025-09-07T08:05:34.2168724Z self.query(hidden_states) 2025-09-07T08:05:34.2168831Z 2025-09-07T08:05:34.2168932Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2169272Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2169576Z return mod(**inputs) 2025-09-07T08:05:34.2169937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2170312Z outputs = self.albert( 2025-09-07T08:05:34.2170664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2171044Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2171413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2171827Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2172239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2172735Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2173235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2173735Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2174228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 385, in forward 2025-09-07T08:05:34.2174607Z self.key(hidden_states) 2025-09-07T08:05:34.2174710Z 2025-09-07T08:05:34.2174803Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2175145Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2175457Z return mod(**inputs) 2025-09-07T08:05:34.2175811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2176177Z outputs = self.albert( 2025-09-07T08:05:34.2176533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2177001Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2177376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2177838Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2178248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2178748Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2179257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2179756Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2180251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 390, in forward 2025-09-07T08:05:34.2180712Z self.value(hidden_states) 2025-09-07T08:05:34.2180823Z 2025-09-07T08:05:34.2180899Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.2181125Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2181466Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2181774Z return mod(**inputs) 2025-09-07T08:05:34.2182116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2182507Z outputs = self.albert( 2025-09-07T08:05:34.2182861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2183241Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2183601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2184024Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2184435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2184934Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2185434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2185923Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2186419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:05:34.2186871Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:34.2187069Z 2025-09-07T08:05:34.2187168Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2187506Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2187811Z return mod(**inputs) 2025-09-07T08:05:34.2188169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2188540Z outputs = self.albert( 2025-09-07T08:05:34.2188891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2189269Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2189639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2190056Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2190465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2191062Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2191560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2192055Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2192544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 415, in forward 2025-09-07T08:05:34.2192954Z projected_context_layer = self.dense(attention_output) 2025-09-07T08:05:34.2193117Z 2025-09-07T08:05:34.2193209Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2193545Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2193853Z return mod(**inputs) 2025-09-07T08:05:34.2194263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2194644Z outputs = self.albert( 2025-09-07T08:05:34.2194998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2195378Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2195751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2196160Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2196571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2197076Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2197581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2197972Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2198360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2198748Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2199129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 462, in ff_chunk 2025-09-07T08:05:34.2199203Z ffn_output = self.ffn(attention_output) 2025-09-07T08:05:34.2199207Z 2025-09-07T08:05:34.2199304Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2199490Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2199551Z return mod(**inputs) 2025-09-07T08:05:34.2199807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2199874Z outputs = self.albert( 2025-09-07T08:05:34.2200128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2200191Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2200446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2200543Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2200800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2200982Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2201232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2201438Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2201694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2201770Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2202025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:05:34.2202102Z ffn_output = self.activation(ffn_output) 2025-09-07T08:05:34.2202317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:34.2202493Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:34.2202497Z 2025-09-07T08:05:34.2202593Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2202781Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2202919Z return mod(**inputs) 2025-09-07T08:05:34.2203334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2203392Z outputs = self.albert( 2025-09-07T08:05:34.2203652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2203717Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2203971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2204074Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2204327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2204516Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2204776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2204855Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2205111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2205178Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2205434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 464, in ff_chunk 2025-09-07T08:05:34.2205506Z ffn_output = self.ffn_output(ffn_output) 2025-09-07T08:05:34.2205510Z 2025-09-07T08:05:34.2205602Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2205788Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2205850Z return mod(**inputs) 2025-09-07T08:05:34.2206109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2206167Z outputs = self.albert( 2025-09-07T08:05:34.2206425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2206488Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2206739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2206835Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2207087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2207269Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2207522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2207882Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2208137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 380, in forward 2025-09-07T08:05:34.2208205Z self.query(hidden_states) 2025-09-07T08:05:34.2208209Z 2025-09-07T08:05:34.2208303Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2208487Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2208544Z return mod(**inputs) 2025-09-07T08:05:34.2208793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2208855Z outputs = self.albert( 2025-09-07T08:05:34.2209214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2209283Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2209533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2209632Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2209883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2210066Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2210317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2210492Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2210745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 385, in forward 2025-09-07T08:05:34.2210812Z self.key(hidden_states) 2025-09-07T08:05:34.2210815Z 2025-09-07T08:05:34.2210906Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2211091Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2211147Z return mod(**inputs) 2025-09-07T08:05:34.2211398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2211456Z outputs = self.albert( 2025-09-07T08:05:34.2211705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2211771Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2212019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2212123Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2212371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2212553Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2212805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2212978Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2213231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 390, in forward 2025-09-07T08:05:34.2213296Z self.value(hidden_states) 2025-09-07T08:05:34.2213300Z 2025-09-07T08:05:34.2213377Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.2213544Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2213728Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2213790Z return mod(**inputs) 2025-09-07T08:05:34.2214043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2214106Z outputs = self.albert( 2025-09-07T08:05:34.2214356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2214420Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2214672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2214767Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2215022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2215279Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2215539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2215712Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2215963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:05:34.2216103Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:34.2216107Z 2025-09-07T08:05:34.2216196Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2216380Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2216436Z return mod(**inputs) 2025-09-07T08:05:34.2216695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2216753Z outputs = self.albert( 2025-09-07T08:05:34.2217002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2217065Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2217313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2217446Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2217697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2217880Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2218135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2218310Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2218562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 415, in forward 2025-09-07T08:05:34.2218663Z projected_context_layer = self.dense(attention_output) 2025-09-07T08:05:34.2218666Z 2025-09-07T08:05:34.2218757Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2218939Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2218995Z return mod(**inputs) 2025-09-07T08:05:34.2219246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2219303Z outputs = self.albert( 2025-09-07T08:05:34.2219559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2219700Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2219950Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2220046Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2220296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2220478Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2220726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2220799Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2221396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2221469Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2221728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 462, in ff_chunk 2025-09-07T08:05:34.2221801Z ffn_output = self.ffn(attention_output) 2025-09-07T08:05:34.2221804Z 2025-09-07T08:05:34.2221896Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2222079Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2222139Z return mod(**inputs) 2025-09-07T08:05:34.2222392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2222450Z outputs = self.albert( 2025-09-07T08:05:34.2222710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2222780Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2223036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2223133Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2223383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2223567Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2223819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2223891Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2224144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2224220Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2224474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:05:34.2224548Z ffn_output = self.activation(ffn_output) 2025-09-07T08:05:34.2224760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:34.2224930Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:34.2224933Z 2025-09-07T08:05:34.2225025Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2225207Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2225263Z return mod(**inputs) 2025-09-07T08:05:34.2225515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2225680Z outputs = self.albert( 2025-09-07T08:05:34.2225934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2225999Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2226256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2226357Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2226610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2226796Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2227045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2227116Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2227438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2227504Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2227760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 464, in ff_chunk 2025-09-07T08:05:34.2227830Z ffn_output = self.ffn_output(ffn_output) 2025-09-07T08:05:34.2227833Z 2025-09-07T08:05:34.2227923Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2228103Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2228161Z return mod(**inputs) 2025-09-07T08:05:34.2228411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2228468Z outputs = self.albert( 2025-09-07T08:05:34.2228853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2228920Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2229175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2229272Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2229521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2229703Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2229951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2230131Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2230386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 380, in forward 2025-09-07T08:05:34.2230450Z self.query(hidden_states) 2025-09-07T08:05:34.2230453Z 2025-09-07T08:05:34.2230540Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2230723Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2230785Z return mod(**inputs) 2025-09-07T08:05:34.2231037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2231103Z outputs = self.albert( 2025-09-07T08:05:34.2231355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2231419Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2231681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2231853Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2232111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2232290Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2232547Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2232723Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2232974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 385, in forward 2025-09-07T08:05:34.2233040Z self.key(hidden_states) 2025-09-07T08:05:34.2233043Z 2025-09-07T08:05:34.2233133Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2233390Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2233451Z return mod(**inputs) 2025-09-07T08:05:34.2233707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2233768Z outputs = self.albert( 2025-09-07T08:05:34.2234017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2234087Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2234336Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2234443Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2234694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2234886Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2235143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2235317Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2235576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 390, in forward 2025-09-07T08:05:34.2235636Z self.value(hidden_states) 2025-09-07T08:05:34.2235639Z 2025-09-07T08:05:34.2235712Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.2235800Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2235982Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2236041Z return mod(**inputs) 2025-09-07T08:05:34.2236296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2236355Z outputs = self.albert( 2025-09-07T08:05:34.2236603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2236664Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2236916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2237013Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2237265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2237444Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2237699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2237949Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2238199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:05:34.2238334Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:34.2238337Z 2025-09-07T08:05:34.2238425Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2238609Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2238665Z return mod(**inputs) 2025-09-07T08:05:34.2238916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2238976Z outputs = self.albert( 2025-09-07T08:05:34.2239294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2239359Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2239610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2239707Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2239958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2240139Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2240390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2240562Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2240818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 415, in forward 2025-09-07T08:05:34.2240918Z projected_context_layer = self.dense(attention_output) 2025-09-07T08:05:34.2240922Z 2025-09-07T08:05:34.2241010Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2241192Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2241247Z return mod(**inputs) 2025-09-07T08:05:34.2241499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2241555Z outputs = self.albert( 2025-09-07T08:05:34.2241805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2241866Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2242117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2242218Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2242466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2242647Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2242897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2242969Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2243224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2243290Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2243550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 462, in ff_chunk 2025-09-07T08:05:34.2243708Z ffn_output = self.ffn(attention_output) 2025-09-07T08:05:34.2243711Z 2025-09-07T08:05:34.2243800Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2243979Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2244036Z return mod(**inputs) 2025-09-07T08:05:34.2244287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2244343Z outputs = self.albert( 2025-09-07T08:05:34.2244595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2244657Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2244906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2245106Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2245359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2245544Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2245797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2245871Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2246125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2246193Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2246454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:05:34.2246534Z ffn_output = self.activation(ffn_output) 2025-09-07T08:05:34.2246751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:34.2246923Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:34.2246926Z 2025-09-07T08:05:34.2247018Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2247202Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2247260Z return mod(**inputs) 2025-09-07T08:05:34.2247513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2247571Z outputs = self.albert( 2025-09-07T08:05:34.2247827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2247895Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2248148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2248252Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2248505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2248691Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2248942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2249017Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2249272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2249337Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2249684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 464, in ff_chunk 2025-09-07T08:05:34.2249757Z ffn_output = self.ffn_output(ffn_output) 2025-09-07T08:05:34.2249760Z 2025-09-07T08:05:34.2249856Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2250044Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2250104Z return mod(**inputs) 2025-09-07T08:05:34.2250358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2250418Z outputs = self.albert( 2025-09-07T08:05:34.2250674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2250737Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2251059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2251168Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2251421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2251614Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2251865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2252049Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2252301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 380, in forward 2025-09-07T08:05:34.2252363Z self.query(hidden_states) 2025-09-07T08:05:34.2252372Z 2025-09-07T08:05:34.2252469Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2252650Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2252707Z return mod(**inputs) 2025-09-07T08:05:34.2252957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2253017Z outputs = self.albert( 2025-09-07T08:05:34.2253267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2253330Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2253583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2253677Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2253929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2254111Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2254360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2254536Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2254784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 385, in forward 2025-09-07T08:05:34.2254846Z self.key(hidden_states) 2025-09-07T08:05:34.2254849Z 2025-09-07T08:05:34.2254938Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2255121Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2255177Z return mod(**inputs) 2025-09-07T08:05:34.2255516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2255576Z outputs = self.albert( 2025-09-07T08:05:34.2255825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2255889Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2256139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2256239Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2256492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2256673Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2256997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2257179Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2257477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 390, in forward 2025-09-07T08:05:34.2257542Z self.value(hidden_states) 2025-09-07T08:05:34.2257545Z 2025-09-07T08:05:34.2257616Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.2257707Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2257891Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2257948Z return mod(**inputs) 2025-09-07T08:05:34.2258198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2258256Z outputs = self.albert( 2025-09-07T08:05:34.2258515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2258577Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2258828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2258924Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2259173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2259353Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2259601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2259776Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2260033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:05:34.2260168Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:34.2260171Z 2025-09-07T08:05:34.2260258Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2260443Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2260499Z return mod(**inputs) 2025-09-07T08:05:34.2260748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2260810Z outputs = self.albert( 2025-09-07T08:05:34.2261060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2261125Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2261375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2261552Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2261804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2261986Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2262238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2262409Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2262662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 415, in forward 2025-09-07T08:05:34.2262760Z projected_context_layer = self.dense(attention_output) 2025-09-07T08:05:34.2262767Z 2025-09-07T08:05:34.2262926Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2263116Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2263172Z return mod(**inputs) 2025-09-07T08:05:34.2263426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2263482Z outputs = self.albert( 2025-09-07T08:05:34.2263734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2263798Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2264046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2264144Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2264394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2264591Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2264840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2264912Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2265172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2265238Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2265497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 462, in ff_chunk 2025-09-07T08:05:34.2265568Z ffn_output = self.ffn(attention_output) 2025-09-07T08:05:34.2265572Z 2025-09-07T08:05:34.2265661Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2265856Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2265912Z return mod(**inputs) 2025-09-07T08:05:34.2266168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2266228Z outputs = self.albert( 2025-09-07T08:05:34.2266485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2266549Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2266798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2266900Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2267151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2267419Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2267672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2267741Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2268002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2268070Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2268330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:05:34.2268405Z ffn_output = self.activation(ffn_output) 2025-09-07T08:05:34.2268614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:34.2268874Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:34.2268881Z 2025-09-07T08:05:34.2268973Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2269160Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2269219Z return mod(**inputs) 2025-09-07T08:05:34.2269471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2269527Z outputs = self.albert( 2025-09-07T08:05:34.2269777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2269858Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2270106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2270209Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2270461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2270647Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2270897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2270965Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2271222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2271290Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2271549Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 464, in ff_chunk 2025-09-07T08:05:34.2271621Z ffn_output = self.ffn_output(ffn_output) 2025-09-07T08:05:34.2271626Z 2025-09-07T08:05:34.2271716Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2271902Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2271959Z return mod(**inputs) 2025-09-07T08:05:34.2272222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2272282Z outputs = self.albert( 2025-09-07T08:05:34.2272540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2272603Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2272853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2272955Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2273206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2273466Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2273716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2273891Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2274143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 380, in forward 2025-09-07T08:05:34.2274208Z self.query(hidden_states) 2025-09-07T08:05:34.2274212Z 2025-09-07T08:05:34.2274303Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2274489Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2274554Z return mod(**inputs) 2025-09-07T08:05:34.2274875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2274937Z outputs = self.albert( 2025-09-07T08:05:34.2275192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2275253Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2275507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2275604Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2275853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2276036Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2276294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2276471Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2276720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 385, in forward 2025-09-07T08:05:34.2276781Z self.key(hidden_states) 2025-09-07T08:05:34.2276785Z 2025-09-07T08:05:34.2276875Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2277057Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2277115Z return mod(**inputs) 2025-09-07T08:05:34.2277364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2277425Z outputs = self.albert( 2025-09-07T08:05:34.2277678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2277746Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2277997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2278092Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2278348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2278529Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2278784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2278955Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2279275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 390, in forward 2025-09-07T08:05:34.2279341Z self.value(hidden_states) 2025-09-07T08:05:34.2279344Z 2025-09-07T08:05:34.2279416Z cudagraph partition due to non gpu ops 2025-09-07T08:05:34.2279510Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2279691Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2279750Z return mod(**inputs) 2025-09-07T08:05:34.2279999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2280058Z outputs = self.albert( 2025-09-07T08:05:34.2280310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2280371Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2280695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2280797Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2281045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2281228Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2281476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2281653Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2281904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 403, in forward 2025-09-07T08:05:34.2282036Z attention_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:05:34.2282043Z 2025-09-07T08:05:34.2282133Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2282317Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2282380Z return mod(**inputs) 2025-09-07T08:05:34.2282629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2282693Z outputs = self.albert( 2025-09-07T08:05:34.2282940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2283002Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2283254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2283351Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2283608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2283788Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2284042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 449, in forward 2025-09-07T08:05:34.2284213Z attention_output = self.attention(hidden_states, attention_mask, head_mask, output_attentions) 2025-09-07T08:05:34.2284463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 415, in forward 2025-09-07T08:05:34.2284568Z projected_context_layer = self.dense(attention_output) 2025-09-07T08:05:34.2284571Z 2025-09-07T08:05:34.2284659Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2284842Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2284977Z return mod(**inputs) 2025-09-07T08:05:34.2285235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2285294Z outputs = self.albert( 2025-09-07T08:05:34.2285542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2285609Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2285858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2285959Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2286210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2286391Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2286717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2286788Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2287043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2287110Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2287371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 462, in ff_chunk 2025-09-07T08:05:34.2287439Z ffn_output = self.ffn(attention_output) 2025-09-07T08:05:34.2287442Z 2025-09-07T08:05:34.2287530Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2287717Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2287773Z return mod(**inputs) 2025-09-07T08:05:34.2288030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2288087Z outputs = self.albert( 2025-09-07T08:05:34.2288337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2288405Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2288653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2288756Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2289005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2289184Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2289439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2289511Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2289764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2289828Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2290084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 463, in ff_chunk 2025-09-07T08:05:34.2290159Z ffn_output = self.activation(ffn_output) 2025-09-07T08:05:34.2290364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:05:34.2290537Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:05:34.2290540Z 2025-09-07T08:05:34.2290627Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2290916Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2290974Z return mod(**inputs) 2025-09-07T08:05:34.2291229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 971, in forward 2025-09-07T08:05:34.2291290Z outputs = self.albert( 2025-09-07T08:05:34.2291537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 722, in forward 2025-09-07T08:05:34.2291606Z encoder_outputs = self.encoder( 2025-09-07T08:05:34.2291856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 534, in forward 2025-09-07T08:05:34.2291959Z layer_group_output = self.albert_layer_groups[group_idx]( 2025-09-07T08:05:34.2292207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 486, in forward 2025-09-07T08:05:34.2292457Z layer_output = albert_layer(hidden_states, attention_mask, head_mask[layer_index], output_attentions) 2025-09-07T08:05:34.2292721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 451, in forward 2025-09-07T08:05:34.2292797Z ffn_output = apply_chunking_to_forward( 2025-09-07T08:05:34.2293056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:05:34.2293122Z return forward_fn(*input_tensors) 2025-09-07T08:05:34.2293380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 464, in ff_chunk 2025-09-07T08:05:34.2293454Z ffn_output = self.ffn_output(ffn_output) 2025-09-07T08:05:34.2293458Z 2025-09-07T08:05:34.2293548Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2293733Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2293794Z return mod(**inputs) 2025-09-07T08:05:34.2294049Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 984, in forward 2025-09-07T08:05:34.2294155Z prediction_scores = self.predictions(sequence_outputs) 2025-09-07T08:05:34.2294405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 866, in forward 2025-09-07T08:05:34.2294486Z hidden_states = self.dense(hidden_states) 2025-09-07T08:05:34.2294489Z 2025-09-07T08:05:34.2294577Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:05:34.2294761Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:05:34.2294818Z return mod(**inputs) 2025-09-07T08:05:34.2295070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 984, in forward 2025-09-07T08:05:34.2295174Z prediction_scores = self.predictions(sequence_outputs) 2025-09-07T08:05:34.2295426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/albert/modeling_albert.py", line 869, in forward 2025-09-07T08:05:34.2295511Z hidden_states = self.decoder(hidden_states) 2025-09-07T08:05:34.2295515Z 2025-09-07T08:05:44.0644991Z pass 2025-09-07T08:05:44.0645373Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:05:46.5245454Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:05:46.5246632Z import pynvml # type: ignore[import] 2025-09-07T08:05:48.5314102Z 2025-09-07T08:05:52.7043802Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:05:52.7044138Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:05:52.7335890Z cpu eval hf_Bart 2025-09-07T08:05:54.0808191Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:05:54.4314342Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:05:54.7777835Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:06:03.9736327Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9736629Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9736852Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9737046Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9737255Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9737449Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9737722Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9738531Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9738892Z return mod(**inputs) 2025-09-07T08:06:03.9739307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9739724Z outputs = self.model( 2025-09-07T08:06:03.9740122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9740527Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9740909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9741331Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9741680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9742049Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9742435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9742848Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9743258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:03.9743724Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:03.9743933Z 2025-09-07T08:06:03.9744039Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9744415Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9744738Z return mod(**inputs) 2025-09-07T08:06:03.9745104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9745483Z outputs = self.model( 2025-09-07T08:06:03.9745831Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9746215Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9746587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9746969Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9747313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9747675Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9748055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9748454Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9748844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:03.9749218Z key_states = self.k_proj(current_states) 2025-09-07T08:06:03.9749503Z 2025-09-07T08:06:03.9749681Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9750035Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9750362Z return mod(**inputs) 2025-09-07T08:06:03.9750724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9751098Z outputs = self.model( 2025-09-07T08:06:03.9751446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9751825Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9752208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9752587Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9752983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9753348Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9753728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9754132Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9754531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:03.9754957Z value_states = self.v_proj(current_states) 2025-09-07T08:06:03.9755098Z 2025-09-07T08:06:03.9755175Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9755378Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9755570Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9755757Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9755977Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9756336Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9756657Z return mod(**inputs) 2025-09-07T08:06:03.9757008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9757388Z outputs = self.model( 2025-09-07T08:06:03.9757750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9758139Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9758527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9758897Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9759250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9759615Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9759999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9760388Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9760784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:03.9761190Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:03.9761639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:03.9762136Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:03.9762320Z 2025-09-07T08:06:03.9762422Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9762788Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9763127Z return mod(**inputs) 2025-09-07T08:06:03.9763574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9763991Z outputs = self.model( 2025-09-07T08:06:03.9764352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9764740Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9765114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9765496Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9765841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9766208Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9766595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9767006Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9767475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:03.9767888Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:03.9768344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:03.9768834Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:03.9768994Z 2025-09-07T08:06:03.9769102Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9769465Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9769782Z return mod(**inputs) 2025-09-07T08:06:03.9770142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9770525Z outputs = self.model( 2025-09-07T08:06:03.9770886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9771275Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9771640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9772022Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9772362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9772728Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9773099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9773510Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9773904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:03.9774289Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:03.9774418Z 2025-09-07T08:06:03.9774522Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9774863Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9775185Z return mod(**inputs) 2025-09-07T08:06:03.9775531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9775910Z outputs = self.model( 2025-09-07T08:06:03.9776257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9776633Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9776992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9777365Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9777904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9778268Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9778669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 323, in forward 2025-09-07T08:06:03.9779100Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:03.9779274Z 2025-09-07T08:06:03.9779381Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9779732Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9780044Z return mod(**inputs) 2025-09-07T08:06:03.9780394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9780779Z outputs = self.model( 2025-09-07T08:06:03.9781196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9781595Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9781955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9782337Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9782677Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9783033Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9783448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 323, in forward 2025-09-07T08:06:03.9783881Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:03.9784259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:03.9784600Z return self.act(input) 2025-09-07T08:06:03.9784706Z 2025-09-07T08:06:03.9784813Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9785149Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9785468Z return mod(**inputs) 2025-09-07T08:06:03.9785817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9786193Z outputs = self.model( 2025-09-07T08:06:03.9786537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9786917Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9787278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9787660Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9788001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9788357Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9788728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 325, in forward 2025-09-07T08:06:03.9789112Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:06:03.9789244Z 2025-09-07T08:06:03.9789348Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9789687Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9790009Z return mod(**inputs) 2025-09-07T08:06:03.9790369Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9790750Z outputs = self.model( 2025-09-07T08:06:03.9791098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9791594Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9791959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9792341Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9792686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9793051Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9793422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9793822Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9794211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:03.9794665Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:03.9794873Z 2025-09-07T08:06:03.9795039Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9795388Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9795705Z return mod(**inputs) 2025-09-07T08:06:03.9796060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9796438Z outputs = self.model( 2025-09-07T08:06:03.9796790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9797162Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9797529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9797904Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9798248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9798608Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9798981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9799373Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9799759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:03.9800141Z key_states = self.k_proj(current_states) 2025-09-07T08:06:03.9800267Z 2025-09-07T08:06:03.9800363Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9800714Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9801046Z return mod(**inputs) 2025-09-07T08:06:03.9801401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9801796Z outputs = self.model( 2025-09-07T08:06:03.9802158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9802541Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9802909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9803558Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9803957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9816206Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9816660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9817087Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9817622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:03.9818223Z value_states = self.v_proj(current_states) 2025-09-07T08:06:03.9818368Z 2025-09-07T08:06:03.9818455Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9818665Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9818855Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9819046Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9819275Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9819646Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9819973Z return mod(**inputs) 2025-09-07T08:06:03.9820367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9820766Z outputs = self.model( 2025-09-07T08:06:03.9821129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9821657Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9822035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9822422Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9822780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9823145Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9823519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9823924Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9824318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:03.9824723Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:03.9825182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:03.9825659Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:03.9825851Z 2025-09-07T08:06:03.9825952Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9826312Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9826636Z return mod(**inputs) 2025-09-07T08:06:03.9826993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9827366Z outputs = self.model( 2025-09-07T08:06:03.9827717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9828100Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9828475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9828856Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9829210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9829572Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9829948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9830345Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9830725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:03.9831118Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:03.9831561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:03.9832103Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:03.9832262Z 2025-09-07T08:06:03.9832363Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9832706Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9833022Z return mod(**inputs) 2025-09-07T08:06:03.9833370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9833745Z outputs = self.model( 2025-09-07T08:06:03.9834090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9834473Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9834836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9835215Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9835629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9835984Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9836386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9836795Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9837199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:03.9837595Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:03.9837730Z 2025-09-07T08:06:03.9837833Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9838194Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9838522Z return mod(**inputs) 2025-09-07T08:06:03.9838893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9839271Z outputs = self.model( 2025-09-07T08:06:03.9839632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9840019Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9840395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9840779Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9841122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9841492Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9841882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 323, in forward 2025-09-07T08:06:03.9842307Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:03.9842486Z 2025-09-07T08:06:03.9842592Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9842933Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9843250Z return mod(**inputs) 2025-09-07T08:06:03.9843604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9843984Z outputs = self.model( 2025-09-07T08:06:03.9844323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9844704Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9845069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9845445Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9845784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9846229Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9846605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 323, in forward 2025-09-07T08:06:03.9847019Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:03.9847399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:03.9847736Z return self.act(input) 2025-09-07T08:06:03.9847839Z 2025-09-07T08:06:03.9847935Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9848277Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9848591Z return mod(**inputs) 2025-09-07T08:06:03.9848940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9850826Z outputs = self.model( 2025-09-07T08:06:03.9851188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9851566Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9851929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9852304Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9852632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9852987Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9853358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 325, in forward 2025-09-07T08:06:03.9853740Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:06:03.9853872Z 2025-09-07T08:06:03.9853971Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9854313Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9854625Z return mod(**inputs) 2025-09-07T08:06:03.9854971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9855346Z outputs = self.model( 2025-09-07T08:06:03.9855688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9856060Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9856421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9856794Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9857124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9857477Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9857922Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9858324Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9858715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:03.9859172Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:03.9859383Z 2025-09-07T08:06:03.9859481Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9859833Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9860155Z return mod(**inputs) 2025-09-07T08:06:03.9860509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9860975Z outputs = self.model( 2025-09-07T08:06:03.9861330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9861709Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9862075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9862457Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9862790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9863145Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9863520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9863915Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9864294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:03.9864745Z key_states = self.k_proj(current_states) 2025-09-07T08:06:03.9864879Z 2025-09-07T08:06:03.9864973Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9865318Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9865634Z return mod(**inputs) 2025-09-07T08:06:03.9865982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9866353Z outputs = self.model( 2025-09-07T08:06:03.9866701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9867080Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9867445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9867820Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9868160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9868513Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9868885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9869268Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9869657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:03.9870048Z value_states = self.v_proj(current_states) 2025-09-07T08:06:03.9870193Z 2025-09-07T08:06:03.9870268Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9870468Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9870665Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9870853Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9871076Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9871425Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9871736Z return mod(**inputs) 2025-09-07T08:06:03.9872084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9872445Z outputs = self.model( 2025-09-07T08:06:03.9872800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9873170Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9873526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9873888Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9874216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9874670Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9875042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9875424Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9875798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:03.9876201Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:03.9876689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:03.9877202Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:03.9877405Z 2025-09-07T08:06:03.9877505Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9877846Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9878256Z return mod(**inputs) 2025-09-07T08:06:03.9878626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9878991Z outputs = self.model( 2025-09-07T08:06:03.9879338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9879703Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9880057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9880421Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9880755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9881095Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9881461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9881851Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9882224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:03.9882610Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:03.9883035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:03.9883482Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:03.9883640Z 2025-09-07T08:06:03.9883734Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9884069Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9884377Z return mod(**inputs) 2025-09-07T08:06:03.9884718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9885084Z outputs = self.model( 2025-09-07T08:06:03.9885428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9885796Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9886149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9886662Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9886993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9887337Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9887708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9888089Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9888552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:03.9888936Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:03.9889072Z 2025-09-07T08:06:03.9889170Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9889529Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9889835Z return mod(**inputs) 2025-09-07T08:06:03.9890190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9890565Z outputs = self.model( 2025-09-07T08:06:03.9890903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9891284Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9891708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9892092Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9892426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9892773Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9893137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 323, in forward 2025-09-07T08:06:03.9893559Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:03.9893731Z 2025-09-07T08:06:03.9893826Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9894162Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9894466Z return mod(**inputs) 2025-09-07T08:06:03.9894805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9895178Z outputs = self.model( 2025-09-07T08:06:03.9895520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9895886Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9896234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9896600Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9896924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9897269Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9897678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 323, in forward 2025-09-07T08:06:03.9898121Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:03.9898512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:03.9898853Z return self.act(input) 2025-09-07T08:06:03.9898959Z 2025-09-07T08:06:03.9899056Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9899397Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9899719Z return mod(**inputs) 2025-09-07T08:06:03.9900080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9900454Z outputs = self.model( 2025-09-07T08:06:03.9900790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9901168Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9901524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9901994Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9902336Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9902694Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9903257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 325, in forward 2025-09-07T08:06:03.9903640Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:06:03.9903770Z 2025-09-07T08:06:03.9903875Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9904224Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9904533Z return mod(**inputs) 2025-09-07T08:06:03.9904888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9905262Z outputs = self.model( 2025-09-07T08:06:03.9905735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9906113Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9906467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9906833Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9907170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9907534Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9907905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9908303Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9908687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:03.9909138Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:03.9909339Z 2025-09-07T08:06:03.9909448Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9909788Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9910108Z return mod(**inputs) 2025-09-07T08:06:03.9910446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9910814Z outputs = self.model( 2025-09-07T08:06:03.9911158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9911525Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9911893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9912258Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9912594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9912947Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9913320Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9913706Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9914084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:03.9914461Z key_states = self.k_proj(current_states) 2025-09-07T08:06:03.9914592Z 2025-09-07T08:06:03.9914683Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9915018Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9915329Z return mod(**inputs) 2025-09-07T08:06:03.9915670Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9916187Z outputs = self.model( 2025-09-07T08:06:03.9916526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9916894Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9917249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9917620Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9917944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9918294Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9918655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9919045Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9919505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:03.9919884Z value_states = self.v_proj(current_states) 2025-09-07T08:06:03.9920018Z 2025-09-07T08:06:03.9920090Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9920291Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9920487Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9920665Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9920880Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9921216Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9921519Z return mod(**inputs) 2025-09-07T08:06:03.9921871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9922239Z outputs = self.model( 2025-09-07T08:06:03.9922580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9922947Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9923317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9923675Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9924010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9924351Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9924727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9925103Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9925480Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:03.9925876Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:03.9926331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:03.9926813Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:03.9926992Z 2025-09-07T08:06:03.9927086Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9927425Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9927738Z return mod(**inputs) 2025-09-07T08:06:03.9928088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9928458Z outputs = self.model( 2025-09-07T08:06:03.9928806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9929282Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9929643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9930017Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9930365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9930715Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9931096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9931481Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9931873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:03.9932262Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:03.9932805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:03.9933269Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:03.9933431Z 2025-09-07T08:06:03.9933526Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9933873Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9934183Z return mod(**inputs) 2025-09-07T08:06:03.9934528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9934909Z outputs = self.model( 2025-09-07T08:06:03.9935259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9935641Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9935993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9936373Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9936713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9937071Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9937435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9937872Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9938259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:03.9938644Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:03.9938773Z 2025-09-07T08:06:03.9938876Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9939210Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9939528Z return mod(**inputs) 2025-09-07T08:06:03.9939881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9940254Z outputs = self.model( 2025-09-07T08:06:03.9940598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9940973Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9941337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9941708Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9942044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9942390Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9942755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 323, in forward 2025-09-07T08:06:03.9943261Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:03.9943430Z 2025-09-07T08:06:03.9943529Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9943869Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9944180Z return mod(**inputs) 2025-09-07T08:06:03.9944526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9944894Z outputs = self.model( 2025-09-07T08:06:03.9945243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9945618Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9945971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9946348Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9946757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9947113Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9947474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 323, in forward 2025-09-07T08:06:03.9947886Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:03.9948265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:03.9948604Z return self.act(input) 2025-09-07T08:06:03.9948708Z 2025-09-07T08:06:03.9948810Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9949148Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9949464Z return mod(**inputs) 2025-09-07T08:06:03.9949817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9950195Z outputs = self.model( 2025-09-07T08:06:03.9950535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9950911Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9951273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9951652Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9951994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9952341Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9952714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 325, in forward 2025-09-07T08:06:03.9953095Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:06:03.9953232Z 2025-09-07T08:06:03.9953335Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9953673Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9953982Z return mod(**inputs) 2025-09-07T08:06:03.9954328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9954699Z outputs = self.model( 2025-09-07T08:06:03.9955042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9955414Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9955779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9956153Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9956490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9956970Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9957354Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9957755Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9958148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:03.9958601Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:03.9958807Z 2025-09-07T08:06:03.9958913Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9959257Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9959574Z return mod(**inputs) 2025-09-07T08:06:03.9959990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9960380Z outputs = self.model( 2025-09-07T08:06:03.9960728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9961109Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9961481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9961861Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9962207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9962560Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9962936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9963328Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9963723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:03.9964098Z key_states = self.k_proj(current_states) 2025-09-07T08:06:03.9964234Z 2025-09-07T08:06:03.9964332Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9964678Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9964995Z return mod(**inputs) 2025-09-07T08:06:03.9965347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9965719Z outputs = self.model( 2025-09-07T08:06:03.9966073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9966450Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9966815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9967192Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9967528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9967881Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9968255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9968652Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9969033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:03.9969415Z value_states = self.v_proj(current_states) 2025-09-07T08:06:03.9969555Z 2025-09-07T08:06:03.9969630Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9969833Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9970029Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9970286Z cudagraph partition due to non gpu ops 2025-09-07T08:06:03.9970504Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9970851Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9971171Z return mod(**inputs) 2025-09-07T08:06:03.9971514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9971889Z outputs = self.model( 2025-09-07T08:06:03.9972240Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9972622Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9972988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9973356Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9973776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9974140Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9974519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9974908Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9975304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:03.9975705Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:03.9976150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:03.9976633Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:03.9976818Z 2025-09-07T08:06:03.9976916Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9977276Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9977660Z return mod(**inputs) 2025-09-07T08:06:03.9978011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9978386Z outputs = self.model( 2025-09-07T08:06:03.9978731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9979108Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9979473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9979853Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9980189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9980549Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9980937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9981332Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9981736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:03.9982128Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:03.9982577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:03.9983034Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:03.9983195Z 2025-09-07T08:06:03.9983300Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9983651Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9984039Z return mod(**inputs) 2025-09-07T08:06:03.9984404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9984783Z outputs = self.model( 2025-09-07T08:06:03.9985135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9985511Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9985879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9986256Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9986595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9986952Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9987321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:03.9987789Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:03.9988181Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:03.9988561Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:03.9988689Z 2025-09-07T08:06:03.9988795Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9989138Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9989458Z return mod(**inputs) 2025-09-07T08:06:03.9989806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9990179Z outputs = self.model( 2025-09-07T08:06:03.9990521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9990898Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9991268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9991644Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9991977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9992342Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9992715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 323, in forward 2025-09-07T08:06:03.9993135Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:03.9993302Z 2025-09-07T08:06:03.9993403Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9993737Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9994052Z return mod(**inputs) 2025-09-07T08:06:03.9994403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:03.9994775Z outputs = self.model( 2025-09-07T08:06:03.9995121Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:03.9995490Z encoder_outputs = self.encoder( 2025-09-07T08:06:03.9995855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:03.9996231Z layer_outputs = encoder_layer( 2025-09-07T08:06:03.9996563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:03.9996914Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:03.9997281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 323, in forward 2025-09-07T08:06:03.9997697Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:03.9998143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:03.9998485Z return self.act(input) 2025-09-07T08:06:03.9998589Z 2025-09-07T08:06:03.9998690Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:03.9999050Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:03.9999366Z return mod(**inputs) 2025-09-07T08:06:03.9999715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0000088Z outputs = self.model( 2025-09-07T08:06:04.0000438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:04.0000812Z encoder_outputs = self.encoder( 2025-09-07T08:06:04.0001240Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:04.0001630Z layer_outputs = encoder_layer( 2025-09-07T08:06:04.0001961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0002316Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0002689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 325, in forward 2025-09-07T08:06:04.0003210Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:06:04.0003344Z 2025-09-07T08:06:04.0003448Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0003828Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0004159Z return mod(**inputs) 2025-09-07T08:06:04.0004525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0004922Z outputs = self.model( 2025-09-07T08:06:04.0005281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:04.0005662Z encoder_outputs = self.encoder( 2025-09-07T08:06:04.0006041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:04.0006423Z layer_outputs = encoder_layer( 2025-09-07T08:06:04.0006770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0007128Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0007512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:04.0007914Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:04.0008309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:04.0008767Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:04.0008968Z 2025-09-07T08:06:04.0009069Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0009426Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0009745Z return mod(**inputs) 2025-09-07T08:06:04.0010100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0010476Z outputs = self.model( 2025-09-07T08:06:04.0010823Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:04.0011208Z encoder_outputs = self.encoder( 2025-09-07T08:06:04.0011581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:04.0012121Z layer_outputs = encoder_layer( 2025-09-07T08:06:04.0012453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0012819Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0013199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:04.0013600Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:04.0013990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:04.0014365Z key_states = self.k_proj(current_states) 2025-09-07T08:06:04.0014508Z 2025-09-07T08:06:04.0014606Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0014958Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0015381Z return mod(**inputs) 2025-09-07T08:06:04.0015735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0016108Z outputs = self.model( 2025-09-07T08:06:04.0016455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:04.0016833Z encoder_outputs = self.encoder( 2025-09-07T08:06:04.0017201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:04.0017612Z layer_outputs = encoder_layer( 2025-09-07T08:06:04.0017957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0018313Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0018695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:04.0019086Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:04.0019464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:04.0019852Z value_states = self.v_proj(current_states) 2025-09-07T08:06:04.0019998Z 2025-09-07T08:06:04.0020078Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0020284Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0020473Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0020670Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0020890Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0021235Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0021548Z return mod(**inputs) 2025-09-07T08:06:04.0021892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0022271Z outputs = self.model( 2025-09-07T08:06:04.0022619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:04.0022999Z encoder_outputs = self.encoder( 2025-09-07T08:06:04.0023358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:04.0023730Z layer_outputs = encoder_layer( 2025-09-07T08:06:04.0024065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0024420Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0024797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:04.0025182Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:04.0025652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0026059Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0026508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:04.0026992Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:04.0027173Z 2025-09-07T08:06:04.0027270Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0027618Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0027938Z return mod(**inputs) 2025-09-07T08:06:04.0028287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0028656Z outputs = self.model( 2025-09-07T08:06:04.0029096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:04.0029484Z encoder_outputs = self.encoder( 2025-09-07T08:06:04.0029848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:04.0030233Z layer_outputs = encoder_layer( 2025-09-07T08:06:04.0030563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0030916Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0031290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:04.0031679Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:04.0032060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0032457Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0032904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:04.0033364Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:04.0033524Z 2025-09-07T08:06:04.0033629Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0033972Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0034292Z return mod(**inputs) 2025-09-07T08:06:04.0034640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0035019Z outputs = self.model( 2025-09-07T08:06:04.0035379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:04.0035757Z encoder_outputs = self.encoder( 2025-09-07T08:06:04.0036132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:04.0036513Z layer_outputs = encoder_layer( 2025-09-07T08:06:04.0036853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0037203Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0037574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 312, in forward 2025-09-07T08:06:04.0037959Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:06:04.0038343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:04.0038722Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:04.0038850Z 2025-09-07T08:06:04.0038949Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0039365Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0039679Z return mod(**inputs) 2025-09-07T08:06:04.0040029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0040405Z outputs = self.model( 2025-09-07T08:06:04.0040745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:04.0041121Z encoder_outputs = self.encoder( 2025-09-07T08:06:04.0041483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:04.0041859Z layer_outputs = encoder_layer( 2025-09-07T08:06:04.0042188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0042541Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0042973Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 323, in forward 2025-09-07T08:06:04.0043405Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:04.0043573Z 2025-09-07T08:06:04.0043674Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0044020Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0044338Z return mod(**inputs) 2025-09-07T08:06:04.0044698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0045075Z outputs = self.model( 2025-09-07T08:06:04.0045422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:04.0045794Z encoder_outputs = self.encoder( 2025-09-07T08:06:04.0046165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:04.0046544Z layer_outputs = encoder_layer( 2025-09-07T08:06:04.0046881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0047226Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0047595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 323, in forward 2025-09-07T08:06:04.0048014Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:04.0048394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:04.0048733Z return self.act(input) 2025-09-07T08:06:04.0048839Z 2025-09-07T08:06:04.0048933Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0049277Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0049598Z return mod(**inputs) 2025-09-07T08:06:04.0049951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0050317Z outputs = self.model( 2025-09-07T08:06:04.0050664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1270, in forward 2025-09-07T08:06:04.0051038Z encoder_outputs = self.encoder( 2025-09-07T08:06:04.0051408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 869, in forward 2025-09-07T08:06:04.0051782Z layer_outputs = encoder_layer( 2025-09-07T08:06:04.0052112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0052471Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0052846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 325, in forward 2025-09-07T08:06:04.0053310Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:06:04.0053441Z 2025-09-07T08:06:04.0053540Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0053875Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0054190Z return mod(**inputs) 2025-09-07T08:06:04.0054535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0054907Z outputs = self.model( 2025-09-07T08:06:04.0055249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0055626Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0055990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0056444Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0056780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0057130Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0057504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0057975Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0058377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:04.0058834Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:04.0059039Z 2025-09-07T08:06:04.0059137Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0059482Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0059802Z return mod(**inputs) 2025-09-07T08:06:04.0060155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0060527Z outputs = self.model( 2025-09-07T08:06:04.0060877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0061255Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0061618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0061993Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0062323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0062673Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0063047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0063456Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0063844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:04.0064232Z key_states = self.k_proj(current_states) 2025-09-07T08:06:04.0064366Z 2025-09-07T08:06:04.0064462Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0064810Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0065126Z return mod(**inputs) 2025-09-07T08:06:04.0065464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0065833Z outputs = self.model( 2025-09-07T08:06:04.0066178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0066639Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0067005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0067371Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0067700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0068041Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0068399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0068782Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0069164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:04.0069545Z value_states = self.v_proj(current_states) 2025-09-07T08:06:04.0069679Z 2025-09-07T08:06:04.0069757Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0070017Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0070207Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0070391Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0070598Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0070931Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0071232Z return mod(**inputs) 2025-09-07T08:06:04.0071569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0071933Z outputs = self.model( 2025-09-07T08:06:04.0072270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0072638Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0072990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0073375Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0073709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0074050Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0074408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0074795Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0075184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0075580Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0076028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:04.0076504Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:04.0076693Z 2025-09-07T08:06:04.0076786Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0077122Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0077438Z return mod(**inputs) 2025-09-07T08:06:04.0077781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0078150Z outputs = self.model( 2025-09-07T08:06:04.0078487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0078858Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0079218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0079722Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0080140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0093527Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0094006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0094462Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0094897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0095325Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0095781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:04.0096244Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:04.0096408Z 2025-09-07T08:06:04.0096510Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0096999Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0097343Z return mod(**inputs) 2025-09-07T08:06:04.0097767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0098158Z outputs = self.model( 2025-09-07T08:06:04.0098508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0098888Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0099256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0099628Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0099970Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0100341Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0100732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0101130Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0101523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:04.0101899Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:04.0102028Z 2025-09-07T08:06:04.0102131Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0102480Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0102794Z return mod(**inputs) 2025-09-07T08:06:04.0103304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0103678Z outputs = self.model( 2025-09-07T08:06:04.0104027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0104401Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0104764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0105133Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0105467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0105814Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0106180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0106587Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0106987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:04.0107625Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:04.0107830Z 2025-09-07T08:06:04.0107932Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0108280Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0108593Z return mod(**inputs) 2025-09-07T08:06:04.0108942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0109317Z outputs = self.model( 2025-09-07T08:06:04.0109664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0110049Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0110417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0110799Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0111235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0111599Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0111980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0112392Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0112800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:04.0113186Z key_states = self.k_proj(current_states) 2025-09-07T08:06:04.0113313Z 2025-09-07T08:06:04.0113413Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0113767Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0114082Z return mod(**inputs) 2025-09-07T08:06:04.0114435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0114802Z outputs = self.model( 2025-09-07T08:06:04.0115151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0115527Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0115893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0116281Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0116614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0116967Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0117341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0117743Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0118155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:04.0118534Z value_states = self.v_proj(current_states) 2025-09-07T08:06:04.0118677Z 2025-09-07T08:06:04.0118755Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0118955Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0119150Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0119334Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0119553Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0119894Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0120214Z return mod(**inputs) 2025-09-07T08:06:04.0120558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0120927Z outputs = self.model( 2025-09-07T08:06:04.0121345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0121724Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0122084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0122455Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0122786Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0123139Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0123507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0123910Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0124300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0124794Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0125246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:04.0125729Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:04.0125910Z 2025-09-07T08:06:04.0126009Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0126348Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0126662Z return mod(**inputs) 2025-09-07T08:06:04.0127006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0127381Z outputs = self.model( 2025-09-07T08:06:04.0127719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0128103Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0128466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0128843Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0129173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0129520Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0129889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0130291Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0130691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0131074Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0131509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:04.0131962Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:04.0132117Z 2025-09-07T08:06:04.0132211Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0132548Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0132851Z return mod(**inputs) 2025-09-07T08:06:04.0133192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0133254Z outputs = self.model( 2025-09-07T08:06:04.0133502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0133567Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0133819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0133956Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0134171Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0134244Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0134488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0134585Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0134826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:04.0134898Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:04.0134903Z 2025-09-07T08:06:04.0134996Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0135184Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0135305Z return mod(**inputs) 2025-09-07T08:06:04.0135557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0135623Z outputs = self.model( 2025-09-07T08:06:04.0135871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0135940Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0136187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0136252Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0136471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0136541Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0136787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 445, in forward 2025-09-07T08:06:04.0136910Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:04.0136914Z 2025-09-07T08:06:04.0137006Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0137196Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0137254Z return mod(**inputs) 2025-09-07T08:06:04.0137499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0137609Z outputs = self.model( 2025-09-07T08:06:04.0137855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0137923Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0138168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0138247Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0138465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0138534Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0138788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 445, in forward 2025-09-07T08:06:04.0138901Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:04.0139119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:04.0139184Z return self.act(input) 2025-09-07T08:06:04.0139187Z 2025-09-07T08:06:04.0139280Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0139474Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0139533Z return mod(**inputs) 2025-09-07T08:06:04.0139859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0139921Z outputs = self.model( 2025-09-07T08:06:04.0140168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0140244Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0140495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0140567Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0140782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0140860Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0141103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 447, in forward 2025-09-07T08:06:04.0141257Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:06:04.0141261Z 2025-09-07T08:06:04.0141362Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0141550Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0141616Z return mod(**inputs) 2025-09-07T08:06:04.0141860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0141923Z outputs = self.model( 2025-09-07T08:06:04.0142175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0142240Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0142491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0142557Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0142779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0142856Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0143097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0143199Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0143443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:04.0143594Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:04.0143597Z 2025-09-07T08:06:04.0143691Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0143879Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0143946Z return mod(**inputs) 2025-09-07T08:06:04.0144196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0144268Z outputs = self.model( 2025-09-07T08:06:04.0144515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0144580Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0144831Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0144895Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0145119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0145188Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0145439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0145592Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0145837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:04.0145917Z key_states = self.k_proj(current_states) 2025-09-07T08:06:04.0145921Z 2025-09-07T08:06:04.0146014Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0146208Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0146266Z return mod(**inputs) 2025-09-07T08:06:04.0146514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0146582Z outputs = self.model( 2025-09-07T08:06:04.0146827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0146901Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0147207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0147274Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0147498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0147569Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0147825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0147909Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0148161Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:04.0148240Z value_states = self.v_proj(current_states) 2025-09-07T08:06:04.0148244Z 2025-09-07T08:06:04.0148321Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0148402Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0148476Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0148555Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0148646Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0148832Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0148898Z return mod(**inputs) 2025-09-07T08:06:04.0149143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0149213Z outputs = self.model( 2025-09-07T08:06:04.0149457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0149522Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0149771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0149839Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0150063Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0150132Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0150380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0150472Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0150715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0150811Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0151109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:04.0151243Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:04.0151303Z 2025-09-07T08:06:04.0151403Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0151597Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0151664Z return mod(**inputs) 2025-09-07T08:06:04.0151919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0151987Z outputs = self.model( 2025-09-07T08:06:04.0152235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0152303Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0152555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0152622Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0152849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0152981Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0153239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0153331Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0153575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0153670Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0153965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:04.0154073Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:04.0154077Z 2025-09-07T08:06:04.0154171Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0154361Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0154436Z return mod(**inputs) 2025-09-07T08:06:04.0154682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0154754Z outputs = self.model( 2025-09-07T08:06:04.0154999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0155071Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0155318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0155383Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0155608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0155683Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0155930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0156019Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0156263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:04.0156334Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:04.0156337Z 2025-09-07T08:06:04.0156436Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0156620Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0156680Z return mod(**inputs) 2025-09-07T08:06:04.0156931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0156992Z outputs = self.model( 2025-09-07T08:06:04.0157244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0157372Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0157625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0157691Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0157905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0157975Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0158217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0158317Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0158569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:04.0158718Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:04.0158725Z 2025-09-07T08:06:04.0158890Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0159093Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0159165Z return mod(**inputs) 2025-09-07T08:06:04.0159420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0159485Z outputs = self.model( 2025-09-07T08:06:04.0159731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0159799Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0160048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0160115Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0160344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0160421Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0160670Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0160780Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0161022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:04.0161098Z key_states = self.k_proj(current_states) 2025-09-07T08:06:04.0161102Z 2025-09-07T08:06:04.0161197Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0161388Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0161452Z return mod(**inputs) 2025-09-07T08:06:04.0161701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0161771Z outputs = self.model( 2025-09-07T08:06:04.0162012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0162082Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0162324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0162389Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0162617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0162696Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0162944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0163042Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0163286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:04.0163431Z value_states = self.v_proj(current_states) 2025-09-07T08:06:04.0163434Z 2025-09-07T08:06:04.0163510Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0163583Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0163652Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0163721Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0163815Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0164003Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0164069Z return mod(**inputs) 2025-09-07T08:06:04.0164313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0164375Z outputs = self.model( 2025-09-07T08:06:04.0164701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0164772Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0165019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0165083Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0165302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0165370Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0165615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0165713Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0165956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0166048Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0166350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:04.0166477Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:04.0166480Z 2025-09-07T08:06:04.0166574Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0166760Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0166821Z return mod(**inputs) 2025-09-07T08:06:04.0167068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0167134Z outputs = self.model( 2025-09-07T08:06:04.0167377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0167442Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0167692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0167758Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0167975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0168042Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0168282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0168379Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0168621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0168707Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0168999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:04.0169444Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:04.0169452Z 2025-09-07T08:06:04.0169545Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0169731Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0169797Z return mod(**inputs) 2025-09-07T08:06:04.0170044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0170108Z outputs = self.model( 2025-09-07T08:06:04.0170350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0170415Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0170663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0170792Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0171014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0171084Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0171328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0171431Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0171670Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:04.0171749Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:04.0171753Z 2025-09-07T08:06:04.0171844Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0172035Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0172096Z return mod(**inputs) 2025-09-07T08:06:04.0172344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0172409Z outputs = self.model( 2025-09-07T08:06:04.0172650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0172720Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0172963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0173030Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0173250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0173319Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0173564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 445, in forward 2025-09-07T08:06:04.0173683Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:04.0173687Z 2025-09-07T08:06:04.0173780Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0173966Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0174025Z return mod(**inputs) 2025-09-07T08:06:04.0174276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0174338Z outputs = self.model( 2025-09-07T08:06:04.0174589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0174654Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0174898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0174966Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0175257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0175334Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0175584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 445, in forward 2025-09-07T08:06:04.0175696Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:04.0175911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:04.0175977Z return self.act(input) 2025-09-07T08:06:04.0175980Z 2025-09-07T08:06:04.0176076Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0176263Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0176323Z return mod(**inputs) 2025-09-07T08:06:04.0176632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0176695Z outputs = self.model( 2025-09-07T08:06:04.0176942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0177007Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0177253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0177319Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0177598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0177674Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0177917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 447, in forward 2025-09-07T08:06:04.0177994Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:06:04.0178001Z 2025-09-07T08:06:04.0178099Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0178284Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0178347Z return mod(**inputs) 2025-09-07T08:06:04.0178590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0178654Z outputs = self.model( 2025-09-07T08:06:04.0178899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0178967Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0179212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0179277Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0179494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0179564Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0179812Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0179903Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0180145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:04.0180288Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:04.0180291Z 2025-09-07T08:06:04.0180382Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0180567Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0180624Z return mod(**inputs) 2025-09-07T08:06:04.0180871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0180996Z outputs = self.model( 2025-09-07T08:06:04.0181241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0181311Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0181554Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0181621Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0181832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0181902Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0182144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0182232Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0182541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:04.0182612Z key_states = self.k_proj(current_states) 2025-09-07T08:06:04.0182616Z 2025-09-07T08:06:04.0182705Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0182893Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0182952Z return mod(**inputs) 2025-09-07T08:06:04.0183198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0183256Z outputs = self.model( 2025-09-07T08:06:04.0183503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0183569Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0183814Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0183883Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0184093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0184161Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0184402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0184485Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0184730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:04.0184805Z value_states = self.v_proj(current_states) 2025-09-07T08:06:04.0184809Z 2025-09-07T08:06:04.0184883Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0184951Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0185020Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0185096Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0185187Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0185375Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0185433Z return mod(**inputs) 2025-09-07T08:06:04.0185676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0185739Z outputs = self.model( 2025-09-07T08:06:04.0185983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0186053Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0186297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0186362Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0186656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0186729Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0186977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0187064Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0187307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0187393Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0187689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:04.0187818Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:04.0187823Z 2025-09-07T08:06:04.0187918Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0188190Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0188249Z return mod(**inputs) 2025-09-07T08:06:04.0188496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0188561Z outputs = self.model( 2025-09-07T08:06:04.0188803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0188873Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0189115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0189181Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0189393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0189460Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0189709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0189793Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0190054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0190139Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0190427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:04.0190529Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:04.0190533Z 2025-09-07T08:06:04.0190622Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0190808Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0190869Z return mod(**inputs) 2025-09-07T08:06:04.0191117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0191175Z outputs = self.model( 2025-09-07T08:06:04.0191417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0191482Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0191724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0191789Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0192001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0192068Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0192311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0192460Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0192706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:04.0192779Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:04.0192782Z 2025-09-07T08:06:04.0192878Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0193065Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0193124Z return mod(**inputs) 2025-09-07T08:06:04.0193376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0193436Z outputs = self.model( 2025-09-07T08:06:04.0193684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0193749Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0194054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0194126Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0194341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0194414Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0194658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0194758Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0195006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:04.0195144Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:04.0195148Z 2025-09-07T08:06:04.0195251Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0195436Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0195501Z return mod(**inputs) 2025-09-07T08:06:04.0195747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0195805Z outputs = self.model( 2025-09-07T08:06:04.0196054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0196119Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0196371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0196435Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0196646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0196721Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0196961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0197060Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0197301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:04.0197372Z key_states = self.k_proj(current_states) 2025-09-07T08:06:04.0197377Z 2025-09-07T08:06:04.0197466Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0197648Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0197709Z return mod(**inputs) 2025-09-07T08:06:04.0197953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0198075Z outputs = self.model( 2025-09-07T08:06:04.0198322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0198387Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0198631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0198693Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0198907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0198974Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0199217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0199317Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0199574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:04.0199716Z value_states = self.v_proj(current_states) 2025-09-07T08:06:04.0199720Z 2025-09-07T08:06:04.0199795Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0199871Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0199939Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0200008Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0200106Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0200298Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0200362Z return mod(**inputs) 2025-09-07T08:06:04.0200609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0200670Z outputs = self.model( 2025-09-07T08:06:04.0200919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0200989Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0201236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0201301Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0201516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0201590Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0201832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0201930Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0202171Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0202257Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0202556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:04.0202679Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:04.0202682Z 2025-09-07T08:06:04.0202779Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0202961Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0203021Z return mod(**inputs) 2025-09-07T08:06:04.0203393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0203458Z outputs = self.model( 2025-09-07T08:06:04.0203711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0203777Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0204030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0204191Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0204408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0204482Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0204729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0204833Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0205076Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0205166Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0205457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:04.0205644Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:04.0205647Z 2025-09-07T08:06:04.0205750Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0205936Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0206002Z return mod(**inputs) 2025-09-07T08:06:04.0206250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0206312Z outputs = self.model( 2025-09-07T08:06:04.0206564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0206632Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0206883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0206950Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0207169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0207242Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0207483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0207580Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0207821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:04.0207898Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:04.0207901Z 2025-09-07T08:06:04.0207993Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0208178Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0208242Z return mod(**inputs) 2025-09-07T08:06:04.0208493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0208559Z outputs = self.model( 2025-09-07T08:06:04.0208804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0208870Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0209115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0209179Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0209397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0209468Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0209709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 445, in forward 2025-09-07T08:06:04.0209821Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:04.0209904Z 2025-09-07T08:06:04.0209995Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0210182Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0210240Z return mod(**inputs) 2025-09-07T08:06:04.0210490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0210550Z outputs = self.model( 2025-09-07T08:06:04.0210795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0210868Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0211116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0211191Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0211461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0211533Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0211787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 445, in forward 2025-09-07T08:06:04.0211891Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:04.0212107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:04.0212171Z return self.act(input) 2025-09-07T08:06:04.0212174Z 2025-09-07T08:06:04.0212271Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0212457Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0212516Z return mod(**inputs) 2025-09-07T08:06:04.0212767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0212833Z outputs = self.model( 2025-09-07T08:06:04.0213082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0213148Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0213394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0213468Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0213681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0213755Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0213997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 447, in forward 2025-09-07T08:06:04.0214072Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:06:04.0214080Z 2025-09-07T08:06:04.0214178Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0214364Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0214431Z return mod(**inputs) 2025-09-07T08:06:04.0214675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0214742Z outputs = self.model( 2025-09-07T08:06:04.0214986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0215052Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0215300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0215367Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0215585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0215718Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0215964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0216058Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0216301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:04.0216449Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:04.0216453Z 2025-09-07T08:06:04.0216544Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0216732Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0216792Z return mod(**inputs) 2025-09-07T08:06:04.0217038Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0217108Z outputs = self.model( 2025-09-07T08:06:04.0217410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0217480Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0217759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0217826Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0218047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0218117Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0218366Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0218456Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0218703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:04.0218783Z key_states = self.k_proj(current_states) 2025-09-07T08:06:04.0218786Z 2025-09-07T08:06:04.0218880Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0219072Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0219131Z return mod(**inputs) 2025-09-07T08:06:04.0219382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0219442Z outputs = self.model( 2025-09-07T08:06:04.0219686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0219759Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0220009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0220089Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0220305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0220377Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0220623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0220709Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0220956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:04.0221033Z value_states = self.v_proj(current_states) 2025-09-07T08:06:04.0221036Z 2025-09-07T08:06:04.0221113Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0221186Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0221255Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0221328Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0221532Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0221720Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0221787Z return mod(**inputs) 2025-09-07T08:06:04.0222038Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0222102Z outputs = self.model( 2025-09-07T08:06:04.0222346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0222419Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0222663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0222728Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0222944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0223070Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0223332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0223423Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0223676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0223770Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0224069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:04.0224198Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:04.0224201Z 2025-09-07T08:06:04.0224297Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0224495Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0224559Z return mod(**inputs) 2025-09-07T08:06:04.0224810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0224875Z outputs = self.model( 2025-09-07T08:06:04.0225120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0225190Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0225433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0225500Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0225721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0225791Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0226046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0226135Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0226379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0226469Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0226758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:04.0226864Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:04.0226868Z 2025-09-07T08:06:04.0226962Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0227157Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0227217Z return mod(**inputs) 2025-09-07T08:06:04.0227470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0227602Z outputs = self.model( 2025-09-07T08:06:04.0227858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0227932Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0228181Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0228248Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0228473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0228547Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0228801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0228894Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0229218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:04.0229297Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:04.0229300Z 2025-09-07T08:06:04.0229396Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0229606Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0229668Z return mod(**inputs) 2025-09-07T08:06:04.0229922Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0229985Z outputs = self.model( 2025-09-07T08:06:04.0230233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0230307Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0230553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0230629Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0230846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0230917Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0231167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0231269Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0231516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:04.0231659Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:04.0231663Z 2025-09-07T08:06:04.0231764Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0231963Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0232023Z return mod(**inputs) 2025-09-07T08:06:04.0232275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0232335Z outputs = self.model( 2025-09-07T08:06:04.0232587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0232652Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0232897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0232968Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0233183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0233260Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0233573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0233680Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0233923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:04.0233998Z key_states = self.k_proj(current_states) 2025-09-07T08:06:04.0234002Z 2025-09-07T08:06:04.0234100Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0234292Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0234356Z return mod(**inputs) 2025-09-07T08:06:04.0234602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0234663Z outputs = self.model( 2025-09-07T08:06:04.0234975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0235044Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0235297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0235362Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0235578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0235655Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0235900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0236000Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0236244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:04.0236334Z value_states = self.v_proj(current_states) 2025-09-07T08:06:04.0236338Z 2025-09-07T08:06:04.0236414Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0236488Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0236564Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0236633Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0236731Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0236915Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0236975Z return mod(**inputs) 2025-09-07T08:06:04.0237226Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0237287Z outputs = self.model( 2025-09-07T08:06:04.0237651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0237718Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0237972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0238045Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0238259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0238334Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0238577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0238672Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0238924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0239017Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0239324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:04.0239513Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:04.0239517Z 2025-09-07T08:06:04.0239615Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0239801Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0239861Z return mod(**inputs) 2025-09-07T08:06:04.0240114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0240174Z outputs = self.model( 2025-09-07T08:06:04.0240428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0240494Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0240740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0240871Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0241094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0241172Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0241418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0241522Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0241766Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0241854Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0242151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:04.0242247Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:04.0242254Z 2025-09-07T08:06:04.0242356Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0242546Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0242611Z return mod(**inputs) 2025-09-07T08:06:04.0242866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0242929Z outputs = self.model( 2025-09-07T08:06:04.0243182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0243248Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0243501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0243569Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0243788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0243867Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0244109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0244209Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0244450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:04.0244523Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:04.0244526Z 2025-09-07T08:06:04.0244621Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0244807Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0244874Z return mod(**inputs) 2025-09-07T08:06:04.0245120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0245449Z outputs = self.model( 2025-09-07T08:06:04.0245701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0245767Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0246019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0246085Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0246304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0246373Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0246619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 445, in forward 2025-09-07T08:06:04.0246739Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:04.0246745Z 2025-09-07T08:06:04.0246895Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0247091Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0247150Z return mod(**inputs) 2025-09-07T08:06:04.0247400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0247467Z outputs = self.model( 2025-09-07T08:06:04.0247714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0247788Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0248033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0248099Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0248338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0248413Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0248670Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 445, in forward 2025-09-07T08:06:04.0248775Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:04.0248992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:04.0249056Z return self.act(input) 2025-09-07T08:06:04.0249060Z 2025-09-07T08:06:04.0249149Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0249339Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0249397Z return mod(**inputs) 2025-09-07T08:06:04.0249650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0249709Z outputs = self.model( 2025-09-07T08:06:04.0249959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0250029Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0250275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0250349Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0250565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0250640Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0250884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 447, in forward 2025-09-07T08:06:04.0250958Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:06:04.0250962Z 2025-09-07T08:06:04.0251059Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0251311Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0251380Z return mod(**inputs) 2025-09-07T08:06:04.0251626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0251687Z outputs = self.model( 2025-09-07T08:06:04.0251940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0252005Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0252252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0252315Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0252531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0252604Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0252922Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0253020Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0253262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:04.0253410Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:04.0253414Z 2025-09-07T08:06:04.0253509Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0253694Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0253755Z return mod(**inputs) 2025-09-07T08:06:04.0253999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0254060Z outputs = self.model( 2025-09-07T08:06:04.0254308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0254372Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0254618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0254683Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0254897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0254965Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0255208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0255297Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0255537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:04.0255619Z key_states = self.k_proj(current_states) 2025-09-07T08:06:04.0255622Z 2025-09-07T08:06:04.0255714Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0255905Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0255965Z return mod(**inputs) 2025-09-07T08:06:04.0256209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0256275Z outputs = self.model( 2025-09-07T08:06:04.0256524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0256592Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0256836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0256903Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0257195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0257267Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0257594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0257683Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0257934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:04.0258013Z value_states = self.v_proj(current_states) 2025-09-07T08:06:04.0258016Z 2025-09-07T08:06:04.0258090Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0258167Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0258237Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0258313Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0258402Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0258654Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0258721Z return mod(**inputs) 2025-09-07T08:06:04.0258968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0259035Z outputs = self.model( 2025-09-07T08:06:04.0259278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0259342Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0259592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0259657Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0259877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0259950Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0260194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0260286Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0260531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0260627Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0260924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:04.0261053Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:04.0261056Z 2025-09-07T08:06:04.0261147Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0261334Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0261404Z return mod(**inputs) 2025-09-07T08:06:04.0261653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0261720Z outputs = self.model( 2025-09-07T08:06:04.0261965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0262029Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0262279Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0262344Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0262562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0262629Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0262879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0263026Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0263271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0263362Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0263652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:04.0263759Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:04.0263763Z 2025-09-07T08:06:04.0263855Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0264042Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0264104Z return mod(**inputs) 2025-09-07T08:06:04.0264349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0264475Z outputs = self.model( 2025-09-07T08:06:04.0264727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0264792Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0265045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0265111Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0265330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0265399Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0265648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0265740Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0265989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:04.0266073Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:04.0266077Z 2025-09-07T08:06:04.0266170Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0266366Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0266428Z return mod(**inputs) 2025-09-07T08:06:04.0266677Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0266744Z outputs = self.model( 2025-09-07T08:06:04.0266993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0267075Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0267320Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0267400Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0267617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0267688Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0267941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0268042Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0268292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:04.0268436Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:04.0268440Z 2025-09-07T08:06:04.0268537Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0268739Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0268862Z return mod(**inputs) 2025-09-07T08:06:04.0269117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0269179Z outputs = self.model( 2025-09-07T08:06:04.0269426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0269506Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0269753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0269830Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0270046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0270123Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0270425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0270530Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0270779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:04.0270854Z key_states = self.k_proj(current_states) 2025-09-07T08:06:04.0270858Z 2025-09-07T08:06:04.0270959Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0271147Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0271209Z return mod(**inputs) 2025-09-07T08:06:04.0271466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0271528Z outputs = self.model( 2025-09-07T08:06:04.0271779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0271851Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0272104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0272171Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0272386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0272463Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0272707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0272809Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0273054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:04.0273132Z value_states = self.v_proj(current_states) 2025-09-07T08:06:04.0273138Z 2025-09-07T08:06:04.0273222Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0273295Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0273372Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0273442Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0273537Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0273730Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0273789Z return mod(**inputs) 2025-09-07T08:06:04.0274042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0274102Z outputs = self.model( 2025-09-07T08:06:04.0274349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0274424Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0274668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0274814Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0275029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0275106Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0275351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0275449Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0275698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0275786Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0276087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:04.0276272Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:04.0276276Z 2025-09-07T08:06:04.0276373Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0276568Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0276629Z return mod(**inputs) 2025-09-07T08:06:04.0276880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0276941Z outputs = self.model( 2025-09-07T08:06:04.0277193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0277259Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0277503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0277577Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0277796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0277875Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0278116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0278212Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0278463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0278547Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0278844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:04.0278942Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:04.0278945Z 2025-09-07T08:06:04.0279050Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0279239Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0279300Z return mod(**inputs) 2025-09-07T08:06:04.0279551Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0279610Z outputs = self.model( 2025-09-07T08:06:04.0279863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0279928Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0280172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0280243Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0280456Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0280592Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0280834Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0280928Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0281178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:04.0281254Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:04.0281257Z 2025-09-07T08:06:04.0281357Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0281541Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0281603Z return mod(**inputs) 2025-09-07T08:06:04.0281847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0281912Z outputs = self.model( 2025-09-07T08:06:04.0282218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0282287Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0282541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0282607Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0282822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0282898Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0283141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 445, in forward 2025-09-07T08:06:04.0283264Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:04.0283268Z 2025-09-07T08:06:04.0283360Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0283552Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0283622Z return mod(**inputs) 2025-09-07T08:06:04.0283864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0283930Z outputs = self.model( 2025-09-07T08:06:04.0284174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0284247Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0284491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0284557Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0284774Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0284843Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0285099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 445, in forward 2025-09-07T08:06:04.0285204Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:04.0285415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:04.0285485Z return self.act(input) 2025-09-07T08:06:04.0285488Z 2025-09-07T08:06:04.0285581Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0285774Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0285832Z return mod(**inputs) 2025-09-07T08:06:04.0286080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0286145Z outputs = self.model( 2025-09-07T08:06:04.0286391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0286523Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0286768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0286840Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0287056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0287125Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0287373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 447, in forward 2025-09-07T08:06:04.0287449Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:06:04.0287452Z 2025-09-07T08:06:04.0287551Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0287738Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0287854Z return mod(**inputs) 2025-09-07T08:06:04.0288111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0288177Z outputs = self.model( 2025-09-07T08:06:04.0288433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0288502Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0288750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0288827Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0289050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0289130Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0289383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0289486Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0289733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:04.0289874Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:04.0289877Z 2025-09-07T08:06:04.0289978Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0290167Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0290235Z return mod(**inputs) 2025-09-07T08:06:04.0290480Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0290545Z outputs = self.model( 2025-09-07T08:06:04.0290797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0290864Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0291118Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0291183Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0291406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0291474Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0291718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0291816Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0292060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:04.0292141Z key_states = self.k_proj(current_states) 2025-09-07T08:06:04.0292202Z 2025-09-07T08:06:04.0292298Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0292489Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0292556Z return mod(**inputs) 2025-09-07T08:06:04.0292804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0292871Z outputs = self.model( 2025-09-07T08:06:04.0293116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0293185Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0293436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0293502Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0293780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0293853Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0294103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0294191Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0294435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:04.0294519Z value_states = self.v_proj(current_states) 2025-09-07T08:06:04.0294522Z 2025-09-07T08:06:04.0294596Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0294671Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0294742Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0294814Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0294912Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0295100Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0295168Z return mod(**inputs) 2025-09-07T08:06:04.0295415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0295475Z outputs = self.model( 2025-09-07T08:06:04.0295727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0295795Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0296047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0296113Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0296329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0296407Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0296657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0296750Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0296992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0297085Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0297379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:04.0297503Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:04.0297506Z 2025-09-07T08:06:04.0297637Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0297823Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0297890Z return mod(**inputs) 2025-09-07T08:06:04.0298222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0298287Z outputs = self.model( 2025-09-07T08:06:04.0298541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0298610Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0298862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0298927Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0299148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0299220Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0299466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0299621Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0299868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0299959Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0300248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:04.0300347Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:04.0300356Z 2025-09-07T08:06:04.0300450Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0300635Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0300702Z return mod(**inputs) 2025-09-07T08:06:04.0300950Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0301018Z outputs = self.model( 2025-09-07T08:06:04.0301265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0301330Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0301580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0301646Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0301865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0301934Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0302176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 413, in forward 2025-09-07T08:06:04.0302270Z hidden_states, self_attn_weights = self.self_attn( 2025-09-07T08:06:04.0302517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:04.0302598Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:04.0302601Z 2025-09-07T08:06:04.0302694Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0302887Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0302948Z return mod(**inputs) 2025-09-07T08:06:04.0303332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0303404Z outputs = self.model( 2025-09-07T08:06:04.0303663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0303736Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0303983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0304147Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0304378Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0304451Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0304709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0304811Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0305056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 216, in forward 2025-09-07T08:06:04.0305208Z query_states = self.q_proj(hidden_states).view(*q_input_shape).transpose(1, 2) 2025-09-07T08:06:04.0305212Z 2025-09-07T08:06:04.0305307Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0305507Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0305571Z return mod(**inputs) 2025-09-07T08:06:04.0305908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0305971Z outputs = self.model( 2025-09-07T08:06:04.0306217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0306289Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0306536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0306608Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0306824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0306895Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0307147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0307249Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0307499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 235, in forward 2025-09-07T08:06:04.0307573Z key_states = self.k_proj(current_states) 2025-09-07T08:06:04.0307577Z 2025-09-07T08:06:04.0307673Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0307860Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0307918Z return mod(**inputs) 2025-09-07T08:06:04.0308172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0308233Z outputs = self.model( 2025-09-07T08:06:04.0308487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0308559Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0308805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0308877Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0309090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0309165Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0309408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0309504Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0309755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 236, in forward 2025-09-07T08:06:04.0309833Z value_states = self.v_proj(current_states) 2025-09-07T08:06:04.0309836Z 2025-09-07T08:06:04.0309916Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0310046Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0310117Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0310189Z cudagraph partition due to non gpu ops 2025-09-07T08:06:04.0310278Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0310469Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0310528Z return mod(**inputs) 2025-09-07T08:06:04.0310774Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0310835Z outputs = self.model( 2025-09-07T08:06:04.0311080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0311149Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0311390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0311519Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0311741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0311815Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0312064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0312164Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0312411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0312498Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0312792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:06:04.0312917Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:04.0312925Z 2025-09-07T08:06:04.0313020Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0313213Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0313271Z return mod(**inputs) 2025-09-07T08:06:04.0313522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0313580Z outputs = self.model( 2025-09-07T08:06:04.0313825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0313892Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0314134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0314201Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0314420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0314492Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0314738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0314836Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0315082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 254, in forward 2025-09-07T08:06:04.0315164Z attn_output, attn_weights = attention_interface( 2025-09-07T08:06:04.0315453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:06:04.0315550Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:06:04.0315554Z 2025-09-07T08:06:04.0315644Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0315933Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0315992Z return mod(**inputs) 2025-09-07T08:06:04.0316241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0316302Z outputs = self.model( 2025-09-07T08:06:04.0316545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0316610Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0316855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0316926Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0317140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0317211Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0317542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 430, in forward 2025-09-07T08:06:04.0317640Z hidden_states, cross_attn_weights = self.encoder_attn( 2025-09-07T08:06:04.0317889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 268, in forward 2025-09-07T08:06:04.0317964Z attn_output = self.out_proj(attn_output) 2025-09-07T08:06:04.0317967Z 2025-09-07T08:06:04.0318062Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0318249Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0318311Z return mod(**inputs) 2025-09-07T08:06:04.0318563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0318623Z outputs = self.model( 2025-09-07T08:06:04.0318914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0318980Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0319225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0319297Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0319510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0319585Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0319826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 445, in forward 2025-09-07T08:06:04.0319942Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:04.0319945Z 2025-09-07T08:06:04.0320038Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0320228Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0320296Z return mod(**inputs) 2025-09-07T08:06:04.0320543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0320609Z outputs = self.model( 2025-09-07T08:06:04.0320851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0320917Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0321168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0321234Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0321453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0321521Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0321766Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 445, in forward 2025-09-07T08:06:04.0321936Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:06:04.0322148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:04.0322220Z return self.act(input) 2025-09-07T08:06:04.0322224Z 2025-09-07T08:06:04.0322313Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0322504Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0322565Z return mod(**inputs) 2025-09-07T08:06:04.0322812Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1471, in forward 2025-09-07T08:06:04.0327928Z outputs = self.model( 2025-09-07T08:06:04.0328255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1288, in forward 2025-09-07T08:06:04.0328433Z decoder_outputs = self.decoder( 2025-09-07T08:06:04.0328725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1115, in forward 2025-09-07T08:06:04.0328798Z layer_outputs = decoder_layer( 2025-09-07T08:06:04.0329045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:04.0329124Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:04.0329387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 447, in forward 2025-09-07T08:06:04.0329465Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:06:04.0329469Z 2025-09-07T08:06:04.0329571Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0329775Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0329843Z return mod(**inputs) 2025-09-07T08:06:04.0330107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1490, in forward 2025-09-07T08:06:04.0330182Z lm_logits = self.lm_head(outputs[0]) 2025-09-07T08:06:04.0330186Z 2025-09-07T08:06:04.0330281Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:04.0330471Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:04.0330529Z return mod(**inputs) 2025-09-07T08:06:04.0330782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bart/modeling_bart.py", line 1491, in forward 2025-09-07T08:06:04.0330905Z lm_logits = lm_logits + self.final_logits_bias.to(lm_logits.device) 2025-09-07T08:06:04.0330909Z 2025-09-07T08:06:16.7290774Z pass 2025-09-07T08:06:16.7291180Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:06:19.4592140Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:06:19.4593046Z import pynvml # type: ignore[import] 2025-09-07T08:06:21.4476638Z 2025-09-07T08:06:24.5647242Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:06:24.5647892Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:06:24.5887106Z cpu eval hf_Bert 2025-09-07T08:06:25.6182284Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:06:25.9062239Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:06:26.1801367Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:06:33.2813419Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2814485Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2814849Z return mod(**inputs) 2025-09-07T08:06:33.2815269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2815674Z outputs = self.bert( 2025-09-07T08:06:33.2816045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 964, in forward 2025-09-07T08:06:33.2816472Z embedding_output = self.embeddings( 2025-09-07T08:06:33.2816951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 188, in forward 2025-09-07T08:06:33.2817348Z embeddings = self.LayerNorm(embeddings) 2025-09-07T08:06:33.2817495Z 2025-09-07T08:06:33.2817614Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.2817826Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.2818169Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.2818364Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.2818564Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.2818758Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.2818954Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.2819143Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.2819338Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.2819550Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.2819766Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.2819951Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.2820187Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2820563Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2820894Z return mod(**inputs) 2025-09-07T08:06:33.2821274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2821667Z outputs = self.bert( 2025-09-07T08:06:33.2822024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2822445Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2822839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2823230Z layer_outputs = layer_module( 2025-09-07T08:06:33.2823597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2823971Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2824364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.2824762Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.2825167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2825561Z return func(*args, **kwargs) 2025-09-07T08:06:33.2825930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.2826365Z self_outputs = self.self( 2025-09-07T08:06:33.2826740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2827112Z return func(*args, **kwargs) 2025-09-07T08:06:33.2827467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:06:33.2828012Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:06:33.2828293Z 2025-09-07T08:06:33.2828399Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2828842Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2829169Z return mod(**inputs) 2025-09-07T08:06:33.2829541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2829923Z outputs = self.bert( 2025-09-07T08:06:33.2830278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2830681Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2831054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2831435Z layer_outputs = layer_module( 2025-09-07T08:06:33.2831798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2832167Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2832640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.2833065Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.2833434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2833805Z return func(*args, **kwargs) 2025-09-07T08:06:33.2834168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.2834547Z self_outputs = self.self( 2025-09-07T08:06:33.2834896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2835265Z return func(*args, **kwargs) 2025-09-07T08:06:33.2835632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:06:33.2836022Z self.key(current_states) 2025-09-07T08:06:33.2836133Z 2025-09-07T08:06:33.2836234Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2836581Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2836905Z return mod(**inputs) 2025-09-07T08:06:33.2837259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2837629Z outputs = self.bert( 2025-09-07T08:06:33.2837970Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2838360Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2838728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2839103Z layer_outputs = layer_module( 2025-09-07T08:06:33.2839446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2839795Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2840174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.2840558Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.2840940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2841308Z return func(*args, **kwargs) 2025-09-07T08:06:33.2841658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.2842030Z self_outputs = self.self( 2025-09-07T08:06:33.2842384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2842835Z return func(*args, **kwargs) 2025-09-07T08:06:33.2843195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:06:33.2843571Z self.value(current_states) 2025-09-07T08:06:33.2843689Z 2025-09-07T08:06:33.2843768Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.2844005Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2844351Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2844658Z return mod(**inputs) 2025-09-07T08:06:33.2845009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2845383Z outputs = self.bert( 2025-09-07T08:06:33.2845726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2846102Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2846535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2846913Z layer_outputs = layer_module( 2025-09-07T08:06:33.2847260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2847627Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2847999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.2848388Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.2848760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2849125Z return func(*args, **kwargs) 2025-09-07T08:06:33.2849478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.2849871Z self_outputs = self.self( 2025-09-07T08:06:33.2850215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2850589Z return func(*args, **kwargs) 2025-09-07T08:06:33.2850941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:06:33.2851376Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:33.2851560Z 2025-09-07T08:06:33.2851664Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2852012Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2852332Z return mod(**inputs) 2025-09-07T08:06:33.2852687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2853069Z outputs = self.bert( 2025-09-07T08:06:33.2853424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2853799Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2854177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2854557Z layer_outputs = layer_module( 2025-09-07T08:06:33.2854898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2855251Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2855633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.2856014Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.2856388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2856840Z return func(*args, **kwargs) 2025-09-07T08:06:33.2857203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:06:33.2857704Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:33.2858132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:06:33.2858521Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.2858657Z 2025-09-07T08:06:33.2858762Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2859114Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2859450Z return mod(**inputs) 2025-09-07T08:06:33.2859815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2860204Z outputs = self.bert( 2025-09-07T08:06:33.2860628Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2861020Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2861400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2861786Z layer_outputs = layer_module( 2025-09-07T08:06:33.2862130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2862481Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2862868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.2863258Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.2863665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.2864065Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.2864477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.2864934Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.2865358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:06:33.2865746Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.2865878Z 2025-09-07T08:06:33.2865979Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2866336Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2866658Z return mod(**inputs) 2025-09-07T08:06:33.2867017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2867398Z outputs = self.bert( 2025-09-07T08:06:33.2867739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2868124Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2868494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2868870Z layer_outputs = layer_module( 2025-09-07T08:06:33.2869203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2869574Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2869948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.2870333Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.2870730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.2871200Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.2871606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.2872062Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.2872482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:06:33.2872897Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:06:33.2873272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:33.2873614Z return self.act(input) 2025-09-07T08:06:33.2873728Z 2025-09-07T08:06:33.2873825Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2874244Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2874562Z return mod(**inputs) 2025-09-07T08:06:33.2874914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2875287Z outputs = self.bert( 2025-09-07T08:06:33.2875635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2876014Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2876377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2876755Z layer_outputs = layer_module( 2025-09-07T08:06:33.2877092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2877447Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2877828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.2878212Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.2878612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.2879004Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.2879409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:06:33.2879869Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:06:33.2880304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:06:33.2880695Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.2880831Z 2025-09-07T08:06:33.2880938Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2881295Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2881606Z return mod(**inputs) 2025-09-07T08:06:33.2881962Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2882335Z outputs = self.bert( 2025-09-07T08:06:33.2882680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2883061Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2883426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2883802Z layer_outputs = layer_module( 2025-09-07T08:06:33.2884142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2884498Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2884942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.2885330Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.2885709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2886086Z return func(*args, **kwargs) 2025-09-07T08:06:33.2886446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.2886814Z self_outputs = self.self( 2025-09-07T08:06:33.2887168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2887536Z return func(*args, **kwargs) 2025-09-07T08:06:33.2887899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:06:33.2888480Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:06:33.2888752Z 2025-09-07T08:06:33.2888855Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2889210Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2889528Z return mod(**inputs) 2025-09-07T08:06:33.2889880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2890250Z outputs = self.bert( 2025-09-07T08:06:33.2890599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2890980Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2891348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2891730Z layer_outputs = layer_module( 2025-09-07T08:06:33.2892066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2892427Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2892801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.2893187Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.2893573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2893952Z return func(*args, **kwargs) 2025-09-07T08:06:33.2894313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.2894694Z self_outputs = self.self( 2025-09-07T08:06:33.2895046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2895414Z return func(*args, **kwargs) 2025-09-07T08:06:33.2895770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:06:33.2896143Z self.key(current_states) 2025-09-07T08:06:33.2896252Z 2025-09-07T08:06:33.2896354Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2896706Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2897018Z return mod(**inputs) 2025-09-07T08:06:33.2897368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2897826Z outputs = self.bert( 2025-09-07T08:06:33.2898176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2898545Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2898989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2899361Z layer_outputs = layer_module( 2025-09-07T08:06:33.2899704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2900280Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2900653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.2901040Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.2901413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2901782Z return func(*args, **kwargs) 2025-09-07T08:06:33.2902136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.2902566Z self_outputs = self.self( 2025-09-07T08:06:33.2902916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2903452Z return func(*args, **kwargs) 2025-09-07T08:06:33.2903824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:06:33.2904198Z self.value(current_states) 2025-09-07T08:06:33.2904317Z 2025-09-07T08:06:33.2904393Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.2904627Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2904990Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2905309Z return mod(**inputs) 2025-09-07T08:06:33.2905663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2906045Z outputs = self.bert( 2025-09-07T08:06:33.2906398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2906777Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2907147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2907529Z layer_outputs = layer_module( 2025-09-07T08:06:33.2907871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2908228Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2908605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.2908987Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.2909362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2909739Z return func(*args, **kwargs) 2025-09-07T08:06:33.2910101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.2910469Z self_outputs = self.self( 2025-09-07T08:06:33.2910817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2911184Z return func(*args, **kwargs) 2025-09-07T08:06:33.2911541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:06:33.2911972Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:33.2912154Z 2025-09-07T08:06:33.2912254Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2912608Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2913085Z return mod(**inputs) 2025-09-07T08:06:33.2913452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2913827Z outputs = self.bert( 2025-09-07T08:06:33.2914169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2914556Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2914923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2915306Z layer_outputs = layer_module( 2025-09-07T08:06:33.2915640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2915998Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2916374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.2917773Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.2918165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2918527Z return func(*args, **kwargs) 2025-09-07T08:06:33.2918890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:06:33.2919318Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:33.2919743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:06:33.2920132Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.2920267Z 2025-09-07T08:06:33.2920368Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2920725Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2921052Z return mod(**inputs) 2025-09-07T08:06:33.2921488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2921865Z outputs = self.bert( 2025-09-07T08:06:33.2922210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2922603Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2922969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2923343Z layer_outputs = layer_module( 2025-09-07T08:06:33.2923675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2924031Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2924402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.2924801Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.2925195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.2925597Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.2926010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.2926473Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.2926919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:06:33.2927315Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.2927461Z 2025-09-07T08:06:33.2927564Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2927922Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2928332Z return mod(**inputs) 2025-09-07T08:06:33.2928697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2929078Z outputs = self.bert( 2025-09-07T08:06:33.2929443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2929842Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2930253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2930631Z layer_outputs = layer_module( 2025-09-07T08:06:33.2930980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2931350Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2931807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.2932217Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.2932619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.2933020Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.2933436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.2933902Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.2934333Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:06:33.2934745Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:06:33.2935129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:33.2935473Z return self.act(input) 2025-09-07T08:06:33.2935582Z 2025-09-07T08:06:33.2935691Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2936052Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2936369Z return mod(**inputs) 2025-09-07T08:06:33.2936728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2937107Z outputs = self.bert( 2025-09-07T08:06:33.2937453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2937903Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2938280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2938661Z layer_outputs = layer_module( 2025-09-07T08:06:33.2939009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2939374Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2939743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.2940134Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.2940544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.2940945Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.2941351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:06:33.2941820Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:06:33.2942258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:06:33.2942731Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.2942864Z 2025-09-07T08:06:33.2942973Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2943320Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2943639Z return mod(**inputs) 2025-09-07T08:06:33.2943993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2944371Z outputs = self.bert( 2025-09-07T08:06:33.2944720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2945100Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2945466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2945844Z layer_outputs = layer_module( 2025-09-07T08:06:33.2946244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2946610Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2947003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.2947396Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.2947782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2948155Z return func(*args, **kwargs) 2025-09-07T08:06:33.2948512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.2948888Z self_outputs = self.self( 2025-09-07T08:06:33.2949247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2949629Z return func(*args, **kwargs) 2025-09-07T08:06:33.2949983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:06:33.2950505Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:06:33.2950777Z 2025-09-07T08:06:33.2950876Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2951229Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2951551Z return mod(**inputs) 2025-09-07T08:06:33.2951897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2952273Z outputs = self.bert( 2025-09-07T08:06:33.2952619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2953012Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2953377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2953747Z layer_outputs = layer_module( 2025-09-07T08:06:33.2954084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2954442Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2954824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.2955207Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.2955576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2955956Z return func(*args, **kwargs) 2025-09-07T08:06:33.2956322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.2956769Z self_outputs = self.self( 2025-09-07T08:06:33.2957117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2957482Z return func(*args, **kwargs) 2025-09-07T08:06:33.2957838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:06:33.2958214Z self.key(current_states) 2025-09-07T08:06:33.2958325Z 2025-09-07T08:06:33.2958427Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2958775Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2959091Z return mod(**inputs) 2025-09-07T08:06:33.2959448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2959834Z outputs = self.bert( 2025-09-07T08:06:33.2960237Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2960624Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2960993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2961372Z layer_outputs = layer_module( 2025-09-07T08:06:33.2961708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2962085Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2962460Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.2962844Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.2963216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2963585Z return func(*args, **kwargs) 2025-09-07T08:06:33.2963939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.2964315Z self_outputs = self.self( 2025-09-07T08:06:33.2964667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2965037Z return func(*args, **kwargs) 2025-09-07T08:06:33.2965385Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:06:33.2965762Z self.value(current_states) 2025-09-07T08:06:33.2965881Z 2025-09-07T08:06:33.2965960Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.2966193Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2966542Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2966859Z return mod(**inputs) 2025-09-07T08:06:33.2967210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2967590Z outputs = self.bert( 2025-09-07T08:06:33.2967935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2968309Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2968675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2969051Z layer_outputs = layer_module( 2025-09-07T08:06:33.2969387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2969739Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2970113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.2970602Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.2970975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2971345Z return func(*args, **kwargs) 2025-09-07T08:06:33.2971693Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.2972069Z self_outputs = self.self( 2025-09-07T08:06:33.2972416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2972785Z return func(*args, **kwargs) 2025-09-07T08:06:33.2973138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:06:33.2973563Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:33.2973757Z 2025-09-07T08:06:33.2973912Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2974263Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2974580Z return mod(**inputs) 2025-09-07T08:06:33.2974931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2975302Z outputs = self.bert( 2025-09-07T08:06:33.2975647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2976026Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2976390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2976762Z layer_outputs = layer_module( 2025-09-07T08:06:33.2977103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2977466Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2977892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.2978276Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.2978641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.2979008Z return func(*args, **kwargs) 2025-09-07T08:06:33.2979361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:06:33.2979795Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:33.2980218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:06:33.2980604Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.2980751Z 2025-09-07T08:06:33.2980851Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2981201Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2981516Z return mod(**inputs) 2025-09-07T08:06:33.2981857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2982228Z outputs = self.bert( 2025-09-07T08:06:33.2982575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2982955Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2983330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2983697Z layer_outputs = layer_module( 2025-09-07T08:06:33.2984037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2984476Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2984853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.2985234Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.2985638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.2986038Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.2986463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.2986925Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.2987345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:06:33.2987738Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.2987935Z 2025-09-07T08:06:33.2988034Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2988392Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2988710Z return mod(**inputs) 2025-09-07T08:06:33.2989056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2989430Z outputs = self.bert( 2025-09-07T08:06:33.2989776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2990164Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2990525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2990903Z layer_outputs = layer_module( 2025-09-07T08:06:33.2991242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2991605Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2991983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.2992364Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.2992765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.2993159Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.2993566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.2994025Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.2994439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:06:33.2994860Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:06:33.2995240Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:33.2995583Z return self.act(input) 2025-09-07T08:06:33.2995688Z 2025-09-07T08:06:33.2995785Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.2996137Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.2996464Z return mod(**inputs) 2025-09-07T08:06:33.2996819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.2997200Z outputs = self.bert( 2025-09-07T08:06:33.2997546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.2997934Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.2998374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.2998755Z layer_outputs = layer_module( 2025-09-07T08:06:33.2999086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.2999445Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.2999821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3000209Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3000609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3001001Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3001408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:06:33.3001984Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:06:33.3002431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:06:33.3002822Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3002952Z 2025-09-07T08:06:33.3003176Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3003546Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3003868Z return mod(**inputs) 2025-09-07T08:06:33.3004220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3004593Z outputs = self.bert( 2025-09-07T08:06:33.3004935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3005319Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3005688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3006066Z layer_outputs = layer_module( 2025-09-07T08:06:33.3006395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3006749Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3007128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3007513Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3007890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3008258Z return func(*args, **kwargs) 2025-09-07T08:06:33.3008618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3009000Z self_outputs = self.self( 2025-09-07T08:06:33.3009351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3009715Z return func(*args, **kwargs) 2025-09-07T08:06:33.3010072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:06:33.3010591Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:06:33.3010861Z 2025-09-07T08:06:33.3010958Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3011307Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3011622Z return mod(**inputs) 2025-09-07T08:06:33.3011976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3012467Z outputs = self.bert( 2025-09-07T08:06:33.3012827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3013209Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3013580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3013962Z layer_outputs = layer_module( 2025-09-07T08:06:33.3014312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3014674Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3015050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3015440Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3015900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3016286Z return func(*args, **kwargs) 2025-09-07T08:06:33.3016647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3017018Z self_outputs = self.self( 2025-09-07T08:06:33.3017370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3017791Z return func(*args, **kwargs) 2025-09-07T08:06:33.3018150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:06:33.3018527Z self.key(current_states) 2025-09-07T08:06:33.3018639Z 2025-09-07T08:06:33.3018737Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3019084Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3019411Z return mod(**inputs) 2025-09-07T08:06:33.3019764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3020135Z outputs = self.bert( 2025-09-07T08:06:33.3020486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3020866Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3021236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3021617Z layer_outputs = layer_module( 2025-09-07T08:06:33.3021951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3022311Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3022684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3023078Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3023443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3023813Z return func(*args, **kwargs) 2025-09-07T08:06:33.3024168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3024543Z self_outputs = self.self( 2025-09-07T08:06:33.3024892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3025253Z return func(*args, **kwargs) 2025-09-07T08:06:33.3025609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:06:33.3025986Z self.value(current_states) 2025-09-07T08:06:33.3026098Z 2025-09-07T08:06:33.3026272Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.3026501Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3026849Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3027165Z return mod(**inputs) 2025-09-07T08:06:33.3027517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3027897Z outputs = self.bert( 2025-09-07T08:06:33.3028235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3028615Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3028983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3029358Z layer_outputs = layer_module( 2025-09-07T08:06:33.3029747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3030116Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3030491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3030880Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3031253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3031618Z return func(*args, **kwargs) 2025-09-07T08:06:33.3031972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3032351Z self_outputs = self.self( 2025-09-07T08:06:33.3032699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3033063Z return func(*args, **kwargs) 2025-09-07T08:06:33.3033418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:06:33.3033845Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:33.3034034Z 2025-09-07T08:06:33.3034130Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3034474Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3034784Z return mod(**inputs) 2025-09-07T08:06:33.3035135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3035510Z outputs = self.bert( 2025-09-07T08:06:33.3035856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3036235Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3036598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3036979Z layer_outputs = layer_module( 2025-09-07T08:06:33.3037313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3037669Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3038043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3038421Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3038797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3039165Z return func(*args, **kwargs) 2025-09-07T08:06:33.3039518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:06:33.3039940Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:33.3040448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:06:33.3040839Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3040974Z 2025-09-07T08:06:33.3041075Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3041417Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3041727Z return mod(**inputs) 2025-09-07T08:06:33.3042075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3042451Z outputs = self.bert( 2025-09-07T08:06:33.3042800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3043184Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3043597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3043994Z layer_outputs = layer_module( 2025-09-07T08:06:33.3044338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3044698Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3045067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3045457Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3045856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3046255Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3046661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3047117Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3047537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:06:33.3047923Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3048053Z 2025-09-07T08:06:33.3048159Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3048505Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3048822Z return mod(**inputs) 2025-09-07T08:06:33.3049173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3049548Z outputs = self.bert( 2025-09-07T08:06:33.3049891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3050267Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3050635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3051019Z layer_outputs = layer_module( 2025-09-07T08:06:33.3064352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3064748Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3065144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3065548Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3065965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3066360Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3066776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3067338Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3067758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:06:33.3068171Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:06:33.3068548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:33.3068884Z return self.act(input) 2025-09-07T08:06:33.3068988Z 2025-09-07T08:06:33.3069089Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3069442Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3069755Z return mod(**inputs) 2025-09-07T08:06:33.3070103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3070471Z outputs = self.bert( 2025-09-07T08:06:33.3070941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3071327Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3071698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3072066Z layer_outputs = layer_module( 2025-09-07T08:06:33.3072402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3072754Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3073122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3073500Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3073892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3074282Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3074680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:06:33.3075140Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:06:33.3075568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:06:33.3075947Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3076079Z 2025-09-07T08:06:33.3076174Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3076527Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3076836Z return mod(**inputs) 2025-09-07T08:06:33.3077179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3077551Z outputs = self.bert( 2025-09-07T08:06:33.3077887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3078256Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3078613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3079097Z layer_outputs = layer_module( 2025-09-07T08:06:33.3079429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3079779Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3080148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3080525Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3080893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3081345Z return func(*args, **kwargs) 2025-09-07T08:06:33.3081698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3082063Z self_outputs = self.self( 2025-09-07T08:06:33.3082411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3082786Z return func(*args, **kwargs) 2025-09-07T08:06:33.3083153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:06:33.3083666Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:06:33.3083930Z 2025-09-07T08:06:33.3084032Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3084443Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3084760Z return mod(**inputs) 2025-09-07T08:06:33.3085116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3085482Z outputs = self.bert( 2025-09-07T08:06:33.3085818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3086191Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3086560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3086927Z layer_outputs = layer_module( 2025-09-07T08:06:33.3087258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3087605Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3087975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3088353Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3088717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3089078Z return func(*args, **kwargs) 2025-09-07T08:06:33.3089422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3089782Z self_outputs = self.self( 2025-09-07T08:06:33.3090124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3090479Z return func(*args, **kwargs) 2025-09-07T08:06:33.3090822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:06:33.3091183Z self.key(current_states) 2025-09-07T08:06:33.3091292Z 2025-09-07T08:06:33.3091390Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3091731Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3092038Z return mod(**inputs) 2025-09-07T08:06:33.3092375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3092733Z outputs = self.bert( 2025-09-07T08:06:33.3093062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3093430Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3093780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3094137Z layer_outputs = layer_module( 2025-09-07T08:06:33.3094465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3094892Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3095268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3095646Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3096032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3096400Z return func(*args, **kwargs) 2025-09-07T08:06:33.3096750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3097112Z self_outputs = self.self( 2025-09-07T08:06:33.3097452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3097878Z return func(*args, **kwargs) 2025-09-07T08:06:33.3098296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:06:33.3098670Z self.value(current_states) 2025-09-07T08:06:33.3098778Z 2025-09-07T08:06:33.3098854Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.3099071Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3099411Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3099714Z return mod(**inputs) 2025-09-07T08:06:33.3100052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3100416Z outputs = self.bert( 2025-09-07T08:06:33.3100755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3101123Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3101485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3101855Z layer_outputs = layer_module( 2025-09-07T08:06:33.3102177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3102540Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3102904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3103428Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3103793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3104149Z return func(*args, **kwargs) 2025-09-07T08:06:33.3104496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3104862Z self_outputs = self.self( 2025-09-07T08:06:33.3105219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3105573Z return func(*args, **kwargs) 2025-09-07T08:06:33.3105915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:06:33.3106335Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:33.3106513Z 2025-09-07T08:06:33.3106608Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3106946Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3107249Z return mod(**inputs) 2025-09-07T08:06:33.3107592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3107955Z outputs = self.bert( 2025-09-07T08:06:33.3108289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3108785Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3109136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3109495Z layer_outputs = layer_module( 2025-09-07T08:06:33.3109817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3110157Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3110514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3110882Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3111246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3111597Z return func(*args, **kwargs) 2025-09-07T08:06:33.3112261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:06:33.3112689Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:33.3113099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:06:33.3113475Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3113607Z 2025-09-07T08:06:33.3113703Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3114044Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3114347Z return mod(**inputs) 2025-09-07T08:06:33.3114688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3115054Z outputs = self.bert( 2025-09-07T08:06:33.3115396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3115767Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3116122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3116483Z layer_outputs = layer_module( 2025-09-07T08:06:33.3116810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3117154Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3117513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3117889Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3118283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3118673Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3119071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3119519Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3119930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:06:33.3120299Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3120428Z 2025-09-07T08:06:33.3120526Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3120861Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3121166Z return mod(**inputs) 2025-09-07T08:06:33.3121505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3121868Z outputs = self.bert( 2025-09-07T08:06:33.3122276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3122644Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3122998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3123362Z layer_outputs = layer_module( 2025-09-07T08:06:33.3123691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3124033Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3124395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3124767Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3125149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3125591Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3125982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3126419Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3126822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:06:33.3127225Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:06:33.3127592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:33.3127918Z return self.act(input) 2025-09-07T08:06:33.3128023Z 2025-09-07T08:06:33.3128114Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3128450Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3128756Z return mod(**inputs) 2025-09-07T08:06:33.3129088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3129454Z outputs = self.bert( 2025-09-07T08:06:33.3129800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3130170Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3130540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3130908Z layer_outputs = layer_module( 2025-09-07T08:06:33.3131239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3131597Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3131976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3132368Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3132763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3133155Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3133557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:06:33.3134020Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:06:33.3134449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:06:33.3134833Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3134970Z 2025-09-07T08:06:33.3135068Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3135420Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3135828Z return mod(**inputs) 2025-09-07T08:06:33.3136177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3136550Z outputs = self.bert( 2025-09-07T08:06:33.3136891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3137263Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3137705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3138073Z layer_outputs = layer_module( 2025-09-07T08:06:33.3138406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3138762Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3139204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3139589Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3139963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3140332Z return func(*args, **kwargs) 2025-09-07T08:06:33.3140688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3141057Z self_outputs = self.self( 2025-09-07T08:06:33.3141404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3141776Z return func(*args, **kwargs) 2025-09-07T08:06:33.3142145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:06:33.3142658Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:06:33.3142923Z 2025-09-07T08:06:33.3143029Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3143372Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3143689Z return mod(**inputs) 2025-09-07T08:06:33.3144042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3144408Z outputs = self.bert( 2025-09-07T08:06:33.3144738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3145112Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3145478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3145853Z layer_outputs = layer_module( 2025-09-07T08:06:33.3146191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3146534Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3146902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3147272Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3147638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3147992Z return func(*args, **kwargs) 2025-09-07T08:06:33.3148343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3148705Z self_outputs = self.self( 2025-09-07T08:06:33.3149050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3149480Z return func(*args, **kwargs) 2025-09-07T08:06:33.3149826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:06:33.3150199Z self.key(current_states) 2025-09-07T08:06:33.3150313Z 2025-09-07T08:06:33.3150409Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3150758Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3151071Z return mod(**inputs) 2025-09-07T08:06:33.3151412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3151779Z outputs = self.bert( 2025-09-07T08:06:33.3152116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3152490Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3152906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3153293Z layer_outputs = layer_module( 2025-09-07T08:06:33.3153625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3153983Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3154353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3154730Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3155107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3155467Z return func(*args, **kwargs) 2025-09-07T08:06:33.3155819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3156187Z self_outputs = self.self( 2025-09-07T08:06:33.3156545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3156906Z return func(*args, **kwargs) 2025-09-07T08:06:33.3157254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:06:33.3157623Z self.value(current_states) 2025-09-07T08:06:33.3157732Z 2025-09-07T08:06:33.3157808Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.3158029Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3158376Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3158688Z return mod(**inputs) 2025-09-07T08:06:33.3159026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3159397Z outputs = self.bert( 2025-09-07T08:06:33.3159741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3160118Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3160477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3160844Z layer_outputs = layer_module( 2025-09-07T08:06:33.3161170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3161519Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3161891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3162268Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3162626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3163064Z return func(*args, **kwargs) 2025-09-07T08:06:33.3163427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3163804Z self_outputs = self.self( 2025-09-07T08:06:33.3164150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3164517Z return func(*args, **kwargs) 2025-09-07T08:06:33.3164869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:06:33.3165297Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:33.3165474Z 2025-09-07T08:06:33.3165581Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3165923Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3166232Z return mod(**inputs) 2025-09-07T08:06:33.3166640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3167016Z outputs = self.bert( 2025-09-07T08:06:33.3167354Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3167734Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3168097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3168466Z layer_outputs = layer_module( 2025-09-07T08:06:33.3168798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3169143Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3169510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3169887Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3170260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3170625Z return func(*args, **kwargs) 2025-09-07T08:06:33.3170979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:06:33.3171406Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:33.3171837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:06:33.3172231Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3172361Z 2025-09-07T08:06:33.3172454Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3172798Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3173113Z return mod(**inputs) 2025-09-07T08:06:33.3173462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3173833Z outputs = self.bert( 2025-09-07T08:06:33.3174169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3174547Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3174908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3175279Z layer_outputs = layer_module( 2025-09-07T08:06:33.3175601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3175947Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3176314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3176776Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3177169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3177619Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3178028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3178483Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3178897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:06:33.3179276Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3179407Z 2025-09-07T08:06:33.3179503Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3179844Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3180153Z return mod(**inputs) 2025-09-07T08:06:33.3180569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3180946Z outputs = self.bert( 2025-09-07T08:06:33.3181285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3181657Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3182018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3182390Z layer_outputs = layer_module( 2025-09-07T08:06:33.3182715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3183062Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3183429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3183817Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3184208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3184588Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3184988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3185433Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3185845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:06:33.3186256Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:06:33.3186622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:33.3186958Z return self.act(input) 2025-09-07T08:06:33.3187067Z 2025-09-07T08:06:33.3187169Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3187519Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3187825Z return mod(**inputs) 2025-09-07T08:06:33.3188170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3188544Z outputs = self.bert( 2025-09-07T08:06:33.3188887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3189262Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3189618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3189989Z layer_outputs = layer_module( 2025-09-07T08:06:33.3190318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3190746Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3191116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3191498Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3191892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3192280Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3192676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:06:33.3193131Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:06:33.3193559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:06:33.3193940Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3194132Z 2025-09-07T08:06:33.3194232Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3194578Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3194884Z return mod(**inputs) 2025-09-07T08:06:33.3195230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3195601Z outputs = self.bert( 2025-09-07T08:06:33.3195941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3196309Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3196670Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3197037Z layer_outputs = layer_module( 2025-09-07T08:06:33.3197379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3197736Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3198126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3198508Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3198885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3199252Z return func(*args, **kwargs) 2025-09-07T08:06:33.3199619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3200006Z self_outputs = self.self( 2025-09-07T08:06:33.3200354Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3200715Z return func(*args, **kwargs) 2025-09-07T08:06:33.3201071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:06:33.3201572Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:06:33.3201841Z 2025-09-07T08:06:33.3201936Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3202274Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3202585Z return mod(**inputs) 2025-09-07T08:06:33.3202944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3203441Z outputs = self.bert( 2025-09-07T08:06:33.3203696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3203762Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3204124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3204192Z layer_outputs = layer_module( 2025-09-07T08:06:33.3204412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3204486Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3204729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3204805Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3205045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3205107Z return func(*args, **kwargs) 2025-09-07T08:06:33.3205348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3205500Z self_outputs = self.self( 2025-09-07T08:06:33.3205753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3205817Z return func(*args, **kwargs) 2025-09-07T08:06:33.3206076Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:06:33.3206139Z self.key(current_states) 2025-09-07T08:06:33.3206143Z 2025-09-07T08:06:33.3206238Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3206438Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3206497Z return mod(**inputs) 2025-09-07T08:06:33.3206747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3206801Z outputs = self.bert( 2025-09-07T08:06:33.3207049Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3207121Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3207361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3207424Z layer_outputs = layer_module( 2025-09-07T08:06:33.3207639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3207713Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3207953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3208024Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3208265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3208324Z return func(*args, **kwargs) 2025-09-07T08:06:33.3208573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3208632Z self_outputs = self.self( 2025-09-07T08:06:33.3208867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3208928Z return func(*args, **kwargs) 2025-09-07T08:06:33.3209172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:06:33.3209241Z self.value(current_states) 2025-09-07T08:06:33.3209244Z 2025-09-07T08:06:33.3209319Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.3209409Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3209603Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3209660Z return mod(**inputs) 2025-09-07T08:06:33.3209981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3210037Z outputs = self.bert( 2025-09-07T08:06:33.3210282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3210352Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3210598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3210662Z layer_outputs = layer_module( 2025-09-07T08:06:33.3210875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3210949Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3211197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3211330Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3211574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3211633Z return func(*args, **kwargs) 2025-09-07T08:06:33.3211880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3211944Z self_outputs = self.self( 2025-09-07T08:06:33.3212179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3212243Z return func(*args, **kwargs) 2025-09-07T08:06:33.3212485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:06:33.3212616Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:33.3212620Z 2025-09-07T08:06:33.3212716Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3212916Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3212976Z return mod(**inputs) 2025-09-07T08:06:33.3213229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3213288Z outputs = self.bert( 2025-09-07T08:06:33.3213531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3213601Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3213842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3213905Z layer_outputs = layer_module( 2025-09-07T08:06:33.3214123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3214196Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3214444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3214511Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3214746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3214813Z return func(*args, **kwargs) 2025-09-07T08:06:33.3215054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:06:33.3215174Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:33.3215414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:06:33.3215496Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3215499Z 2025-09-07T08:06:33.3215657Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3215843Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3215903Z return mod(**inputs) 2025-09-07T08:06:33.3216152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3216214Z outputs = self.bert( 2025-09-07T08:06:33.3216459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3216521Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3216767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3216829Z layer_outputs = layer_module( 2025-09-07T08:06:33.3217044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3217181Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3217432Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3217560Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3217825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3217897Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3218175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3218292Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3218536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:06:33.3218610Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3218617Z 2025-09-07T08:06:33.3218713Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3218900Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3218961Z return mod(**inputs) 2025-09-07T08:06:33.3219207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3219261Z outputs = self.bert( 2025-09-07T08:06:33.3219508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3219570Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3219816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3219875Z layer_outputs = layer_module( 2025-09-07T08:06:33.3220092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3220166Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3220406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3220482Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3220739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3220810Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3221083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3221190Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3221432Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:06:33.3221535Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:06:33.3221824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:33.3221886Z return self.act(input) 2025-09-07T08:06:33.3221891Z 2025-09-07T08:06:33.3221986Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3222171Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3222228Z return mod(**inputs) 2025-09-07T08:06:33.3222472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3222529Z outputs = self.bert( 2025-09-07T08:06:33.3222777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3222844Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3223139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3223208Z layer_outputs = layer_module( 2025-09-07T08:06:33.3223423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3223491Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3223729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3223801Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3224064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3224129Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3224406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:06:33.3224535Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:06:33.3224788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:06:33.3224860Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3224863Z 2025-09-07T08:06:33.3224952Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3225143Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3225202Z return mod(**inputs) 2025-09-07T08:06:33.3225457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3225516Z outputs = self.bert( 2025-09-07T08:06:33.3225758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3225829Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3226074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3226141Z layer_outputs = layer_module( 2025-09-07T08:06:33.3226354Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3226423Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3226679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3226751Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3226998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3227058Z return func(*args, **kwargs) 2025-09-07T08:06:33.3227309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3227435Z self_outputs = self.self( 2025-09-07T08:06:33.3227676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3227745Z return func(*args, **kwargs) 2025-09-07T08:06:33.3227985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:06:33.3228196Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:06:33.3228200Z 2025-09-07T08:06:33.3228289Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3228479Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3228544Z return mod(**inputs) 2025-09-07T08:06:33.3228792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3228856Z outputs = self.bert( 2025-09-07T08:06:33.3229157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3229233Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3229479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3229543Z layer_outputs = layer_module( 2025-09-07T08:06:33.3229772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3229843Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3230090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3230160Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3230401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3230477Z return func(*args, **kwargs) 2025-09-07T08:06:33.3230719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3230790Z self_outputs = self.self( 2025-09-07T08:06:33.3231028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3231092Z return func(*args, **kwargs) 2025-09-07T08:06:33.3231342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:06:33.3231406Z self.key(current_states) 2025-09-07T08:06:33.3231409Z 2025-09-07T08:06:33.3231512Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3231700Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3231765Z return mod(**inputs) 2025-09-07T08:06:33.3232018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3232077Z outputs = self.bert( 2025-09-07T08:06:33.3232329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3232395Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3232641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3232703Z layer_outputs = layer_module( 2025-09-07T08:06:33.3232919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3232994Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3233235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3233375Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3233620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3233682Z return func(*args, **kwargs) 2025-09-07T08:06:33.3233932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3233996Z self_outputs = self.self( 2025-09-07T08:06:33.3234239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3234299Z return func(*args, **kwargs) 2025-09-07T08:06:33.3234539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:06:33.3234611Z self.value(current_states) 2025-09-07T08:06:33.3234614Z 2025-09-07T08:06:33.3234690Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.3234851Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3235040Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3235108Z return mod(**inputs) 2025-09-07T08:06:33.3235355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3235416Z outputs = self.bert( 2025-09-07T08:06:33.3235666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3235730Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3235976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3236039Z layer_outputs = layer_module( 2025-09-07T08:06:33.3236252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3236333Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3236574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3236652Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3236891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3236951Z return func(*args, **kwargs) 2025-09-07T08:06:33.3237200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3237260Z self_outputs = self.self( 2025-09-07T08:06:33.3237502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3237562Z return func(*args, **kwargs) 2025-09-07T08:06:33.3237812Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:06:33.3237939Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:33.3237943Z 2025-09-07T08:06:33.3238033Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3238226Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3238281Z return mod(**inputs) 2025-09-07T08:06:33.3238533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3238588Z outputs = self.bert( 2025-09-07T08:06:33.3238835Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3238904Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3239147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3239304Z layer_outputs = layer_module( 2025-09-07T08:06:33.3239521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3239589Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3239837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3239911Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3240155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3240218Z return func(*args, **kwargs) 2025-09-07T08:06:33.3240466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:06:33.3240586Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:33.3240942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:06:33.3241030Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3241033Z 2025-09-07T08:06:33.3241125Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3241318Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3241375Z return mod(**inputs) 2025-09-07T08:06:33.3241623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3241688Z outputs = self.bert( 2025-09-07T08:06:33.3241932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3242003Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3242244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3242318Z layer_outputs = layer_module( 2025-09-07T08:06:33.3242532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3242599Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3242847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3242919Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3243183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3243252Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3243532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3243649Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3243900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:06:33.3243980Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3243983Z 2025-09-07T08:06:33.3244076Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3244265Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3244322Z return mod(**inputs) 2025-09-07T08:06:33.3244566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3244630Z outputs = self.bert( 2025-09-07T08:06:33.3244876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3244944Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3245186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3245311Z layer_outputs = layer_module( 2025-09-07T08:06:33.3245536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3245605Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3245850Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3245922Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3246178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3246249Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3246522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3246653Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3246963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:06:33.3247074Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:06:33.3247286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:33.3247352Z return self.act(input) 2025-09-07T08:06:33.3247355Z 2025-09-07T08:06:33.3247454Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3247637Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3247699Z return mod(**inputs) 2025-09-07T08:06:33.3247945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3248002Z outputs = self.bert( 2025-09-07T08:06:33.3248252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3248319Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3248561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3248620Z layer_outputs = layer_module( 2025-09-07T08:06:33.3248829Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3248896Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3249135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3249211Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3249473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3249543Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3249823Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:06:33.3249949Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:06:33.3250200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:06:33.3250274Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3250277Z 2025-09-07T08:06:33.3250374Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3250565Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3250622Z return mod(**inputs) 2025-09-07T08:06:33.3250873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3250930Z outputs = self.bert( 2025-09-07T08:06:33.3251239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3251302Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3251546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3251606Z layer_outputs = layer_module( 2025-09-07T08:06:33.3251818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3251891Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3252130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3252202Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3252438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3252560Z return func(*args, **kwargs) 2025-09-07T08:06:33.3252804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3252868Z self_outputs = self.self( 2025-09-07T08:06:33.3253104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3253163Z return func(*args, **kwargs) 2025-09-07T08:06:33.3253401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:06:33.3253606Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:06:33.3253611Z 2025-09-07T08:06:33.3253702Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3253887Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3253945Z return mod(**inputs) 2025-09-07T08:06:33.3254197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3254255Z outputs = self.bert( 2025-09-07T08:06:33.3254498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3254568Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3254810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3254880Z layer_outputs = layer_module( 2025-09-07T08:06:33.3255092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3255158Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3255406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3255482Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3255724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3255784Z return func(*args, **kwargs) 2025-09-07T08:06:33.3256026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3256094Z self_outputs = self.self( 2025-09-07T08:06:33.3256332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3256397Z return func(*args, **kwargs) 2025-09-07T08:06:33.3256639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:06:33.3256707Z self.key(current_states) 2025-09-07T08:06:33.3256710Z 2025-09-07T08:06:33.3256801Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3257058Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3257129Z return mod(**inputs) 2025-09-07T08:06:33.3257379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3257441Z outputs = self.bert( 2025-09-07T08:06:33.3257739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3257811Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3258066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3258131Z layer_outputs = layer_module( 2025-09-07T08:06:33.3258356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3258501Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3258750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3258832Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3259075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3259146Z return func(*args, **kwargs) 2025-09-07T08:06:33.3259389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3259459Z self_outputs = self.self( 2025-09-07T08:06:33.3259696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3259759Z return func(*args, **kwargs) 2025-09-07T08:06:33.3260013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:06:33.3260083Z self.value(current_states) 2025-09-07T08:06:33.3260087Z 2025-09-07T08:06:33.3260167Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.3260261Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3260448Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3260513Z return mod(**inputs) 2025-09-07T08:06:33.3260761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3260831Z outputs = self.bert( 2025-09-07T08:06:33.3261077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3261143Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3261390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3261460Z layer_outputs = layer_module( 2025-09-07T08:06:33.3261684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3261754Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3262000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3262070Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3262307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3262375Z return func(*args, **kwargs) 2025-09-07T08:06:33.3262614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3262683Z self_outputs = self.self( 2025-09-07T08:06:33.3262920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3263060Z return func(*args, **kwargs) 2025-09-07T08:06:33.3263311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:06:33.3263440Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:33.3263444Z 2025-09-07T08:06:33.3263542Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3263726Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3263787Z return mod(**inputs) 2025-09-07T08:06:33.3264037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3264098Z outputs = self.bert( 2025-09-07T08:06:33.3264353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3264482Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3264743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3264811Z layer_outputs = layer_module( 2025-09-07T08:06:33.3265027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3265110Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3265353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3265435Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3265673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3265732Z return func(*args, **kwargs) 2025-09-07T08:06:33.3265988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:06:33.3266116Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:33.3266365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:06:33.3266438Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3266441Z 2025-09-07T08:06:33.3266539Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3266724Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3266829Z return mod(**inputs) 2025-09-07T08:06:33.3267244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3267332Z outputs = self.bert( 2025-09-07T08:06:33.3267657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3267756Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3268029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3268182Z layer_outputs = layer_module( 2025-09-07T08:06:33.3268439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3268563Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3268851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3268952Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3269258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3269393Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3269820Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3269969Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3270313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:06:33.3270413Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3270417Z 2025-09-07T08:06:33.3270520Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3270817Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3270917Z return mod(**inputs) 2025-09-07T08:06:33.3271226Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3271311Z outputs = self.bert( 2025-09-07T08:06:33.3271641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3271793Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3272098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3272225Z layer_outputs = layer_module( 2025-09-07T08:06:33.3272475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3272599Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3272858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3273010Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3273341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3273434Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3273773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3273905Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3274248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:06:33.3274396Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:06:33.3274640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:33.3274801Z return self.act(input) 2025-09-07T08:06:33.3274805Z 2025-09-07T08:06:33.3274922Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3275166Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3275296Z return mod(**inputs) 2025-09-07T08:06:33.3275592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3275710Z outputs = self.bert( 2025-09-07T08:06:33.3275984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3276118Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3276379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3276566Z layer_outputs = layer_module( 2025-09-07T08:06:33.3276813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3276909Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3277210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3277386Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3277736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3277844Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3278145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:06:33.3278342Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:06:33.3278612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:06:33.3278730Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3278733Z 2025-09-07T08:06:33.3278892Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3279221Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3279308Z return mod(**inputs) 2025-09-07T08:06:33.3279653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3279783Z outputs = self.bert( 2025-09-07T08:06:33.3280055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3280229Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3280514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3280634Z layer_outputs = layer_module( 2025-09-07T08:06:33.3280875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3280970Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3281294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3281414Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3281736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3281825Z return func(*args, **kwargs) 2025-09-07T08:06:33.3282095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3282204Z self_outputs = self.self( 2025-09-07T08:06:33.3282510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3282656Z return func(*args, **kwargs) 2025-09-07T08:06:33.3282926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:06:33.3283186Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:06:33.3283192Z 2025-09-07T08:06:33.3283313Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3283514Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3283740Z return mod(**inputs) 2025-09-07T08:06:33.3284018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3284130Z outputs = self.bert( 2025-09-07T08:06:33.3284403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3284496Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3284828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3284945Z layer_outputs = layer_module( 2025-09-07T08:06:33.3285220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3285381Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3285684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3285784Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3286097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3286230Z return func(*args, **kwargs) 2025-09-07T08:06:33.3286504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3286623Z self_outputs = self.self( 2025-09-07T08:06:33.3286902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3286980Z return func(*args, **kwargs) 2025-09-07T08:06:33.3287456Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:06:33.3287552Z self.key(current_states) 2025-09-07T08:06:33.3287555Z 2025-09-07T08:06:33.3287706Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3287988Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3288106Z return mod(**inputs) 2025-09-07T08:06:33.3288425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3288523Z outputs = self.bert( 2025-09-07T08:06:33.3288830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3288937Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3289248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3289333Z layer_outputs = layer_module( 2025-09-07T08:06:33.3289624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3289767Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3290042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3290185Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3290459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3290607Z return func(*args, **kwargs) 2025-09-07T08:06:33.3290890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3290981Z self_outputs = self.self( 2025-09-07T08:06:33.3291297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3291390Z return func(*args, **kwargs) 2025-09-07T08:06:33.3291679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:06:33.3291834Z self.value(current_states) 2025-09-07T08:06:33.3291838Z 2025-09-07T08:06:33.3291980Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.3292121Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3292342Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3292511Z return mod(**inputs) 2025-09-07T08:06:33.3292778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3292945Z outputs = self.bert( 2025-09-07T08:06:33.3293235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3293397Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3293703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3293798Z layer_outputs = layer_module( 2025-09-07T08:06:33.3294099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3294221Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3294494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3294623Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3294891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3294997Z return func(*args, **kwargs) 2025-09-07T08:06:33.3295382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3295521Z self_outputs = self.self( 2025-09-07T08:06:33.3295784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3295874Z return func(*args, **kwargs) 2025-09-07T08:06:33.3296173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:06:33.3296329Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:33.3296333Z 2025-09-07T08:06:33.3296536Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3296808Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3296893Z return mod(**inputs) 2025-09-07T08:06:33.3297208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3297309Z outputs = self.bert( 2025-09-07T08:06:33.3297695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3297800Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3298111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3298202Z layer_outputs = layer_module( 2025-09-07T08:06:33.3298448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3298576Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3298883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3299027Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3299296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3299386Z return func(*args, **kwargs) 2025-09-07T08:06:33.3299703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:06:33.3299837Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:33.3300186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:06:33.3300292Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3300296Z 2025-09-07T08:06:33.3300452Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3300681Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3300765Z return mod(**inputs) 2025-09-07T08:06:33.3301103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3301312Z outputs = self.bert( 2025-09-07T08:06:33.3301633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3301727Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3302028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3302106Z layer_outputs = layer_module( 2025-09-07T08:06:33.3302390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3302544Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3302833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3302961Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3303472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3303559Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3303967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3304114Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3304423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:06:33.3304527Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3304531Z 2025-09-07T08:06:33.3304673Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3304934Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3305043Z return mod(**inputs) 2025-09-07T08:06:33.3305365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3305452Z outputs = self.bert( 2025-09-07T08:06:33.3305758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3305840Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3306264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3306361Z layer_outputs = layer_module( 2025-09-07T08:06:33.3306610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3306737Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3307016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3307218Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3307527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3307625Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3307964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3308106Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3308396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:06:33.3308579Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:06:33.3308862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:33.3308953Z return self.act(input) 2025-09-07T08:06:33.3308957Z 2025-09-07T08:06:33.3309083Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3309447Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3309523Z return mod(**inputs) 2025-09-07T08:06:33.3309901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3309994Z outputs = self.bert( 2025-09-07T08:06:33.3310272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3310409Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3310731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3310877Z layer_outputs = layer_module( 2025-09-07T08:06:33.3311138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3311324Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3311618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3311724Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3312035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3312167Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3312518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:06:33.3312683Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:06:33.3312956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:06:33.3313086Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3313092Z 2025-09-07T08:06:33.3313204Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3313503Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3313588Z return mod(**inputs) 2025-09-07T08:06:33.3313905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3313991Z outputs = self.bert( 2025-09-07T08:06:33.3314263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3314409Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3314694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3314827Z layer_outputs = layer_module( 2025-09-07T08:06:33.3315123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3315254Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3315512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3315648Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3315984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3316080Z return func(*args, **kwargs) 2025-09-07T08:06:33.3316377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3316470Z self_outputs = self.self( 2025-09-07T08:06:33.3316723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3316909Z return func(*args, **kwargs) 2025-09-07T08:06:33.3317184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:06:33.3317514Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:06:33.3317517Z 2025-09-07T08:06:33.3317643Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3317880Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3318015Z return mod(**inputs) 2025-09-07T08:06:33.3318309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3318427Z outputs = self.bert( 2025-09-07T08:06:33.3318702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3318823Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3319159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3319292Z layer_outputs = layer_module( 2025-09-07T08:06:33.3319622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3319722Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3320024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3320141Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3320469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3320574Z return func(*args, **kwargs) 2025-09-07T08:06:33.3320844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3320966Z self_outputs = self.self( 2025-09-07T08:06:33.3321249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3321357Z return func(*args, **kwargs) 2025-09-07T08:06:33.3321665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:06:33.3321768Z self.key(current_states) 2025-09-07T08:06:33.3321803Z 2025-09-07T08:06:33.3321926Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3322143Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3322273Z return mod(**inputs) 2025-09-07T08:06:33.3322539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3322708Z outputs = self.bert( 2025-09-07T08:06:33.3322986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3323084Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3323400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3323494Z layer_outputs = layer_module( 2025-09-07T08:06:33.3323798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3323948Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3324219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3324361Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3324628Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3324735Z return func(*args, **kwargs) 2025-09-07T08:06:33.3325034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3325238Z self_outputs = self.self( 2025-09-07T08:06:33.3325505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3325590Z return func(*args, **kwargs) 2025-09-07T08:06:33.3325891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:06:33.3331907Z self.value(current_states) 2025-09-07T08:06:33.3331917Z 2025-09-07T08:06:33.3332019Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.3332130Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3332344Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3332406Z return mod(**inputs) 2025-09-07T08:06:33.3332797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3332866Z outputs = self.bert( 2025-09-07T08:06:33.3333139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3333207Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3333459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3333524Z layer_outputs = layer_module( 2025-09-07T08:06:33.3333743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3333821Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3334069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3334146Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3334409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3334472Z return func(*args, **kwargs) 2025-09-07T08:06:33.3334720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3334783Z self_outputs = self.self( 2025-09-07T08:06:33.3335021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3335079Z return func(*args, **kwargs) 2025-09-07T08:06:33.3335317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:06:33.3335450Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:33.3335454Z 2025-09-07T08:06:33.3335557Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3335771Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3335832Z return mod(**inputs) 2025-09-07T08:06:33.3336085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3336142Z outputs = self.bert( 2025-09-07T08:06:33.3336385Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3336453Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3336691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3336752Z layer_outputs = layer_module( 2025-09-07T08:06:33.3336979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3337048Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3337657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3337731Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3337967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3338027Z return func(*args, **kwargs) 2025-09-07T08:06:33.3338270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:06:33.3338393Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:33.3338634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:06:33.3338713Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3338717Z 2025-09-07T08:06:33.3338813Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3339082Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3339148Z return mod(**inputs) 2025-09-07T08:06:33.3339393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3339455Z outputs = self.bert( 2025-09-07T08:06:33.3339709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3339778Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3340036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3340100Z layer_outputs = layer_module( 2025-09-07T08:06:33.3340327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3340395Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3340646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3340718Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3340983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3341055Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3341334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3341453Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3341702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:06:33.3341781Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3341789Z 2025-09-07T08:06:33.3341887Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3342090Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3342153Z return mod(**inputs) 2025-09-07T08:06:33.3342405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3342468Z outputs = self.bert( 2025-09-07T08:06:33.3342712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3342779Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3343027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3343092Z layer_outputs = layer_module( 2025-09-07T08:06:33.3343312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3343446Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3343692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3343769Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3344029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3344099Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3344376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3344490Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3344733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:06:33.3344835Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:06:33.3345106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:33.3345171Z return self.act(input) 2025-09-07T08:06:33.3345175Z 2025-09-07T08:06:33.3345272Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3345465Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3345520Z return mod(**inputs) 2025-09-07T08:06:33.3345770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3345826Z outputs = self.bert( 2025-09-07T08:06:33.3346081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3346146Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3346384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3346456Z layer_outputs = layer_module( 2025-09-07T08:06:33.3346681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3346751Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3346995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3347070Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3347326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3347391Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3347666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:06:33.3347790Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:06:33.3348046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:06:33.3348117Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3348120Z 2025-09-07T08:06:33.3348212Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3348404Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3348459Z return mod(**inputs) 2025-09-07T08:06:33.3348709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3348764Z outputs = self.bert( 2025-09-07T08:06:33.3349009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3349073Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3349314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3349447Z layer_outputs = layer_module( 2025-09-07T08:06:33.3349663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3349736Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3349992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3350064Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3350313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3350374Z return func(*args, **kwargs) 2025-09-07T08:06:33.3350618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3350680Z self_outputs = self.self( 2025-09-07T08:06:33.3350975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3351043Z return func(*args, **kwargs) 2025-09-07T08:06:33.3351281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:06:33.3351494Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:06:33.3351498Z 2025-09-07T08:06:33.3351586Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3351777Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3351834Z return mod(**inputs) 2025-09-07T08:06:33.3352085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3352146Z outputs = self.bert( 2025-09-07T08:06:33.3352391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3352462Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3352701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3352761Z layer_outputs = layer_module( 2025-09-07T08:06:33.3352980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3353045Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3353289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3353357Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3353589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3353654Z return func(*args, **kwargs) 2025-09-07T08:06:33.3353899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3353963Z self_outputs = self.self( 2025-09-07T08:06:33.3354198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3354262Z return func(*args, **kwargs) 2025-09-07T08:06:33.3354502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:06:33.3354562Z self.key(current_states) 2025-09-07T08:06:33.3354565Z 2025-09-07T08:06:33.3354658Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3354841Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3354900Z return mod(**inputs) 2025-09-07T08:06:33.3355147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3355268Z outputs = self.bert( 2025-09-07T08:06:33.3355515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3355577Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3355822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3355886Z layer_outputs = layer_module( 2025-09-07T08:06:33.3356101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3356174Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3356414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3356490Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3356781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3356848Z return func(*args, **kwargs) 2025-09-07T08:06:33.3357089Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3357152Z self_outputs = self.self( 2025-09-07T08:06:33.3357396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3357456Z return func(*args, **kwargs) 2025-09-07T08:06:33.3357704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:06:33.3357767Z self.value(current_states) 2025-09-07T08:06:33.3357770Z 2025-09-07T08:06:33.3357844Z cudagraph partition due to non gpu ops 2025-09-07T08:06:33.3357941Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3358132Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3358190Z return mod(**inputs) 2025-09-07T08:06:33.3358433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3358490Z outputs = self.bert( 2025-09-07T08:06:33.3358737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3358800Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3359045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3359106Z layer_outputs = layer_module( 2025-09-07T08:06:33.3359322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3359386Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3359632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3359710Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3359946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3360007Z return func(*args, **kwargs) 2025-09-07T08:06:33.3360249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:06:33.3360309Z self_outputs = self.self( 2025-09-07T08:06:33.3360548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3360605Z return func(*args, **kwargs) 2025-09-07T08:06:33.3360845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:06:33.3361053Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:06:33.3361057Z 2025-09-07T08:06:33.3361153Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3361338Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3361393Z return mod(**inputs) 2025-09-07T08:06:33.3361641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3361698Z outputs = self.bert( 2025-09-07T08:06:33.3361946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3362009Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3362250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3362318Z layer_outputs = layer_module( 2025-09-07T08:06:33.3362592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3362664Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3362901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:06:33.3362970Z self_attention_outputs = self.attention( 2025-09-07T08:06:33.3363212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:06:33.3363271Z return func(*args, **kwargs) 2025-09-07T08:06:33.3363511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:06:33.3363628Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:06:33.3363868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:06:33.3363951Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3363954Z 2025-09-07T08:06:33.3364043Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3364237Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3364293Z return mod(**inputs) 2025-09-07T08:06:33.3364539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3364594Z outputs = self.bert( 2025-09-07T08:06:33.3364836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3364897Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3365135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3365194Z layer_outputs = layer_module( 2025-09-07T08:06:33.3365410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3365478Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3365725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3365797Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3366065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3366132Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3366416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3366527Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3366774Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:06:33.3366915Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3366918Z 2025-09-07T08:06:33.3367011Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3367199Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3367254Z return mod(**inputs) 2025-09-07T08:06:33.3367499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3367561Z outputs = self.bert( 2025-09-07T08:06:33.3367802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3367870Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3368110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3368231Z layer_outputs = layer_module( 2025-09-07T08:06:33.3368455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3368520Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3368769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3368839Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3369100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3369163Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3369433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:06:33.3369537Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:06:33.3369782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:06:33.3369885Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:06:33.3370092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:06:33.3370157Z return self.act(input) 2025-09-07T08:06:33.3370162Z 2025-09-07T08:06:33.3370252Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3370435Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3370492Z return mod(**inputs) 2025-09-07T08:06:33.3370735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:06:33.3370791Z outputs = self.bert( 2025-09-07T08:06:33.3371032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:06:33.3371097Z encoder_outputs = self.encoder( 2025-09-07T08:06:33.3371337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:06:33.3371395Z layer_outputs = layer_module( 2025-09-07T08:06:33.3371607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:06:33.3371683Z return super().__call__(*args, **kwargs) 2025-09-07T08:06:33.3371928Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:06:33.3372000Z layer_output = apply_chunking_to_forward( 2025-09-07T08:06:33.3372253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:06:33.3372317Z return forward_fn(*input_tensors) 2025-09-07T08:06:33.3372588Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:06:33.3372775Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:06:33.3373020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:06:33.3373090Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3373093Z 2025-09-07T08:06:33.3373189Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3373376Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3373438Z return mod(**inputs) 2025-09-07T08:06:33.3373684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1323, in forward 2025-09-07T08:06:33.3373770Z prediction_scores = self.cls(sequence_output) 2025-09-07T08:06:33.3374072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 780, in forward 2025-09-07T08:06:33.3374182Z prediction_scores = self.predictions(sequence_output) 2025-09-07T08:06:33.3374432Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 769, in forward 2025-09-07T08:06:33.3374513Z hidden_states = self.transform(hidden_states) 2025-09-07T08:06:33.3374754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 745, in forward 2025-09-07T08:06:33.3374831Z hidden_states = self.dense(hidden_states) 2025-09-07T08:06:33.3374834Z 2025-09-07T08:06:33.3374923Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:06:33.3375111Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:06:33.3375169Z return mod(**inputs) 2025-09-07T08:06:33.3375422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1323, in forward 2025-09-07T08:06:33.3375505Z prediction_scores = self.cls(sequence_output) 2025-09-07T08:06:33.3375746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 780, in forward 2025-09-07T08:06:33.3375843Z prediction_scores = self.predictions(sequence_output) 2025-09-07T08:06:33.3376086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 770, in forward 2025-09-07T08:06:33.3376171Z hidden_states = self.decoder(hidden_states) 2025-09-07T08:06:33.3376175Z 2025-09-07T08:06:41.7343677Z pass 2025-09-07T08:06:41.7344063Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:06:44.3344452Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:06:44.3345457Z import pynvml # type: ignore[import] 2025-09-07T08:06:46.3398394Z 2025-09-07T08:06:51.7048172Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:06:51.7048445Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:06:51.7441845Z cpu eval hf_Bert_large 2025-09-07T08:06:54.0732257Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:06:54.7221263Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:06:55.3606887Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:07:07.1989380Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.1989846Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.1990613Z return mod(**inputs) 2025-09-07T08:07:07.1991028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.1991438Z outputs = self.bert( 2025-09-07T08:07:07.1991813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 964, in forward 2025-09-07T08:07:07.1992239Z embedding_output = self.embeddings( 2025-09-07T08:07:07.1992638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 188, in forward 2025-09-07T08:07:07.1993029Z embeddings = self.LayerNorm(embeddings) 2025-09-07T08:07:07.1993165Z 2025-09-07T08:07:07.1993244Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1993454Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1993647Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1993841Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1994036Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1994377Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1994577Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1994774Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1994980Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1995167Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1995361Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1995593Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1995815Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1996002Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1996191Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1996380Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1996570Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1996754Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1996943Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1997145Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1997338Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1997523Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1997715Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1997904Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.1998140Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.1998516Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.1998846Z return mod(**inputs) 2025-09-07T08:07:07.1999212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.1999592Z outputs = self.bert( 2025-09-07T08:07:07.1999943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2000390Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2000780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2001212Z layer_outputs = layer_module( 2025-09-07T08:07:07.2001584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2001997Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2002404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2002802Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2003323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2003705Z return func(*args, **kwargs) 2025-09-07T08:07:07.2004076Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2004592Z self_outputs = self.self( 2025-09-07T08:07:07.2004964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2005339Z return func(*args, **kwargs) 2025-09-07T08:07:07.2005712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2006239Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2006511Z 2025-09-07T08:07:07.2006619Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2006991Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2007322Z return mod(**inputs) 2025-09-07T08:07:07.2007686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2008058Z outputs = self.bert( 2025-09-07T08:07:07.2008515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2008915Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2009287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2009663Z layer_outputs = layer_module( 2025-09-07T08:07:07.2010017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2010388Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2010767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2011149Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2011520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2011907Z return func(*args, **kwargs) 2025-09-07T08:07:07.2012269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2012645Z self_outputs = self.self( 2025-09-07T08:07:07.2013000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2013392Z return func(*args, **kwargs) 2025-09-07T08:07:07.2013779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2014152Z self.key(current_states) 2025-09-07T08:07:07.2014259Z 2025-09-07T08:07:07.2014370Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2014714Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2015039Z return mod(**inputs) 2025-09-07T08:07:07.2015397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2015768Z outputs = self.bert( 2025-09-07T08:07:07.2016116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2016499Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2016872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2017250Z layer_outputs = layer_module( 2025-09-07T08:07:07.2017646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2018000Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2018384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2018850Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2019240Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2019625Z return func(*args, **kwargs) 2025-09-07T08:07:07.2019990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2020366Z self_outputs = self.self( 2025-09-07T08:07:07.2020714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2021088Z return func(*args, **kwargs) 2025-09-07T08:07:07.2021457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2021831Z self.value(current_states) 2025-09-07T08:07:07.2021951Z 2025-09-07T08:07:07.2022026Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2022258Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2022681Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2023000Z return mod(**inputs) 2025-09-07T08:07:07.2023363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2023749Z outputs = self.bert( 2025-09-07T08:07:07.2024106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2024508Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2024881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2025267Z layer_outputs = layer_module( 2025-09-07T08:07:07.2025620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2025996Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2026377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2026759Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2027141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2027512Z return func(*args, **kwargs) 2025-09-07T08:07:07.2027876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2028247Z self_outputs = self.self( 2025-09-07T08:07:07.2028599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2028963Z return func(*args, **kwargs) 2025-09-07T08:07:07.2029322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2029762Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2029954Z 2025-09-07T08:07:07.2030055Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2030432Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2030760Z return mod(**inputs) 2025-09-07T08:07:07.2031125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2031514Z outputs = self.bert( 2025-09-07T08:07:07.2031856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2032247Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2032616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2033110Z layer_outputs = layer_module( 2025-09-07T08:07:07.2033443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2033809Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2034189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2034579Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2034960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2035322Z return func(*args, **kwargs) 2025-09-07T08:07:07.2035726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2036169Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2036703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2037109Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2037245Z 2025-09-07T08:07:07.2037345Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2037705Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2038025Z return mod(**inputs) 2025-09-07T08:07:07.2038388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2038764Z outputs = self.bert( 2025-09-07T08:07:07.2039109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2039493Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2039871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2040261Z layer_outputs = layer_module( 2025-09-07T08:07:07.2040596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2040953Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2041342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2041761Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2042170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2042558Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2042978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2043478Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2043927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2044310Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2044451Z 2025-09-07T08:07:07.2044554Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2044908Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2045231Z return mod(**inputs) 2025-09-07T08:07:07.2045580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2045951Z outputs = self.bert( 2025-09-07T08:07:07.2046303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2046686Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2047058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2047511Z layer_outputs = layer_module( 2025-09-07T08:07:07.2047856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2048219Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2048603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2048985Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2049373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2049786Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2050192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2050650Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2051133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2051555Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2051942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2052282Z return self.act(input) 2025-09-07T08:07:07.2052393Z 2025-09-07T08:07:07.2052500Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2052858Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2053169Z return mod(**inputs) 2025-09-07T08:07:07.2053517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2053901Z outputs = self.bert( 2025-09-07T08:07:07.2054256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2054631Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2055005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2055387Z layer_outputs = layer_module( 2025-09-07T08:07:07.2055729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2056085Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2056452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2056843Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2057237Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2057694Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2058110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2058568Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2059022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2059413Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2059552Z 2025-09-07T08:07:07.2059654Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2060006Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2060327Z return mod(**inputs) 2025-09-07T08:07:07.2060695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2061099Z outputs = self.bert( 2025-09-07T08:07:07.2061533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2061909Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2062279Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2062657Z layer_outputs = layer_module( 2025-09-07T08:07:07.2062998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2063352Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2063724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2064105Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2064492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2064935Z return func(*args, **kwargs) 2025-09-07T08:07:07.2065300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2065670Z self_outputs = self.self( 2025-09-07T08:07:07.2066033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2066403Z return func(*args, **kwargs) 2025-09-07T08:07:07.2066762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2067277Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2067546Z 2025-09-07T08:07:07.2067640Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2067991Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2068314Z return mod(**inputs) 2025-09-07T08:07:07.2068667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2069035Z outputs = self.bert( 2025-09-07T08:07:07.2069375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2069755Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2070121Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2070492Z layer_outputs = layer_module( 2025-09-07T08:07:07.2070819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2071171Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2071544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2071948Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2072310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2072677Z return func(*args, **kwargs) 2025-09-07T08:07:07.2073033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2073427Z self_outputs = self.self( 2025-09-07T08:07:07.2073777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2074137Z return func(*args, **kwargs) 2025-09-07T08:07:07.2074489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2074864Z self.key(current_states) 2025-09-07T08:07:07.2074974Z 2025-09-07T08:07:07.2075078Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2075506Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2075816Z return mod(**inputs) 2025-09-07T08:07:07.2076162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2076535Z outputs = self.bert( 2025-09-07T08:07:07.2076881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2077254Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2077621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2077994Z layer_outputs = layer_module( 2025-09-07T08:07:07.2078327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2078742Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2079108Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2079487Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2079857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2080223Z return func(*args, **kwargs) 2025-09-07T08:07:07.2080569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2080939Z self_outputs = self.self( 2025-09-07T08:07:07.2081285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2081646Z return func(*args, **kwargs) 2025-09-07T08:07:07.2081998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2082371Z self.value(current_states) 2025-09-07T08:07:07.2082490Z 2025-09-07T08:07:07.2082566Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2082789Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2083144Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2083458Z return mod(**inputs) 2025-09-07T08:07:07.2083803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2084176Z outputs = self.bert( 2025-09-07T08:07:07.2084522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2084913Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2085268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2085652Z layer_outputs = layer_module( 2025-09-07T08:07:07.2085982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2086330Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2086696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2087076Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2087443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2087806Z return func(*args, **kwargs) 2025-09-07T08:07:07.2088159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2088524Z self_outputs = self.self( 2025-09-07T08:07:07.2088872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2090440Z return func(*args, **kwargs) 2025-09-07T08:07:07.2090801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2091244Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2091426Z 2025-09-07T08:07:07.2091521Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2091873Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2092188Z return mod(**inputs) 2025-09-07T08:07:07.2092537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2092905Z outputs = self.bert( 2025-09-07T08:07:07.2093245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2093692Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2094058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2094434Z layer_outputs = layer_module( 2025-09-07T08:07:07.2094759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2095114Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2095484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2095863Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2096237Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2096602Z return func(*args, **kwargs) 2025-09-07T08:07:07.2096959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2097390Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2097873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2098252Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2098391Z 2025-09-07T08:07:07.2098482Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2098823Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2099138Z return mod(**inputs) 2025-09-07T08:07:07.2099480Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2099848Z outputs = self.bert( 2025-09-07T08:07:07.2100189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2100570Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2100933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2101299Z layer_outputs = layer_module( 2025-09-07T08:07:07.2101635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2101983Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2102357Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2102741Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2103317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2103719Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2104248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2104707Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2105130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2105507Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2105643Z 2025-09-07T08:07:07.2105737Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2106083Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2106396Z return mod(**inputs) 2025-09-07T08:07:07.2106735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2107111Z outputs = self.bert( 2025-09-07T08:07:07.2107561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2107966Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2108346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2108725Z layer_outputs = layer_module( 2025-09-07T08:07:07.2109064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2109423Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2109803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2110191Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2110589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2110983Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2111398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2111857Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2112268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2112686Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2113064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2113406Z return self.act(input) 2025-09-07T08:07:07.2113957Z 2025-09-07T08:07:07.2114064Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2114412Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2114733Z return mod(**inputs) 2025-09-07T08:07:07.2115094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2115473Z outputs = self.bert( 2025-09-07T08:07:07.2115825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2116199Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2116567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2116947Z layer_outputs = layer_module( 2025-09-07T08:07:07.2117290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2117640Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2118016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2118401Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2118889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2119285Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2119682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2120142Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2120578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2120963Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2121095Z 2025-09-07T08:07:07.2121198Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2121540Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2121864Z return mod(**inputs) 2025-09-07T08:07:07.2122277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2122657Z outputs = self.bert( 2025-09-07T08:07:07.2122997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2123376Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2123742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2124117Z layer_outputs = layer_module( 2025-09-07T08:07:07.2124454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2124807Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2125194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2125585Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2125962Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2126331Z return func(*args, **kwargs) 2025-09-07T08:07:07.2126686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2127060Z self_outputs = self.self( 2025-09-07T08:07:07.2127412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2127786Z return func(*args, **kwargs) 2025-09-07T08:07:07.2128134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2128670Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2128942Z 2025-09-07T08:07:07.2129047Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2129401Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2129720Z return mod(**inputs) 2025-09-07T08:07:07.2130078Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2130453Z outputs = self.bert( 2025-09-07T08:07:07.2130801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2131181Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2131547Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2131915Z layer_outputs = layer_module( 2025-09-07T08:07:07.2132252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2132692Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2133080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2133465Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2133847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2134217Z return func(*args, **kwargs) 2025-09-07T08:07:07.2134585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2134959Z self_outputs = self.self( 2025-09-07T08:07:07.2135307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2135670Z return func(*args, **kwargs) 2025-09-07T08:07:07.2136099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2136482Z self.key(current_states) 2025-09-07T08:07:07.2136591Z 2025-09-07T08:07:07.2136698Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2137044Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2137362Z return mod(**inputs) 2025-09-07T08:07:07.2137795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2138182Z outputs = self.bert( 2025-09-07T08:07:07.2138525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2138907Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2139278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2139659Z layer_outputs = layer_module( 2025-09-07T08:07:07.2139996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2140350Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2140725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2141109Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2141479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2141848Z return func(*args, **kwargs) 2025-09-07T08:07:07.2142206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2142579Z self_outputs = self.self( 2025-09-07T08:07:07.2142934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2143302Z return func(*args, **kwargs) 2025-09-07T08:07:07.2143647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2144015Z self.value(current_states) 2025-09-07T08:07:07.2144129Z 2025-09-07T08:07:07.2144205Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2144433Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2144778Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2145097Z return mod(**inputs) 2025-09-07T08:07:07.2145450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2145831Z outputs = self.bert( 2025-09-07T08:07:07.2146173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2146639Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2147005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2147382Z layer_outputs = layer_module( 2025-09-07T08:07:07.2147716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2148063Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2148438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2148819Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2149190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2149553Z return func(*args, **kwargs) 2025-09-07T08:07:07.2149963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2150350Z self_outputs = self.self( 2025-09-07T08:07:07.2150706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2151077Z return func(*args, **kwargs) 2025-09-07T08:07:07.2151439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2151866Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2152057Z 2025-09-07T08:07:07.2152156Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2152509Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2152831Z return mod(**inputs) 2025-09-07T08:07:07.2153182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2153656Z outputs = self.bert( 2025-09-07T08:07:07.2154002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2154382Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2154749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2155118Z layer_outputs = layer_module( 2025-09-07T08:07:07.2155459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2155816Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2156192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2156576Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2156948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2157318Z return func(*args, **kwargs) 2025-09-07T08:07:07.2157676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2158107Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2158523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2158911Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2159054Z 2025-09-07T08:07:07.2159148Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2159498Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2159814Z return mod(**inputs) 2025-09-07T08:07:07.2160162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2160614Z outputs = self.bert( 2025-09-07T08:07:07.2160960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2161340Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2161699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2162077Z layer_outputs = layer_module( 2025-09-07T08:07:07.2162414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2162772Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2163377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2163881Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2164438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2176456Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2176988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2177496Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2178025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2178424Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2178571Z 2025-09-07T08:07:07.2178676Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2179045Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2179377Z return mod(**inputs) 2025-09-07T08:07:07.2179750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2180139Z outputs = self.bert( 2025-09-07T08:07:07.2180490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2180873Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2181257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2181633Z layer_outputs = layer_module( 2025-09-07T08:07:07.2181979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2182345Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2182730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2183120Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2183522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2183919Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2184329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2184790Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2185208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2185624Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2186007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2186374Z return self.act(input) 2025-09-07T08:07:07.2186482Z 2025-09-07T08:07:07.2186592Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2187067Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2187390Z return mod(**inputs) 2025-09-07T08:07:07.2187754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2188135Z outputs = self.bert( 2025-09-07T08:07:07.2188495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2188874Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2189252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2189636Z layer_outputs = layer_module( 2025-09-07T08:07:07.2189981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2190340Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2190791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2191190Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2191625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2192025Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2192430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2192901Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2193344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2193735Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2193876Z 2025-09-07T08:07:07.2193989Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2194342Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2194666Z return mod(**inputs) 2025-09-07T08:07:07.2195024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2195403Z outputs = self.bert( 2025-09-07T08:07:07.2195740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2196121Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2196488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2196863Z layer_outputs = layer_module( 2025-09-07T08:07:07.2197204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2197578Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2197955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2198354Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2198735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2199103Z return func(*args, **kwargs) 2025-09-07T08:07:07.2199462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2199834Z self_outputs = self.self( 2025-09-07T08:07:07.2200194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2200562Z return func(*args, **kwargs) 2025-09-07T08:07:07.2200921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2201520Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2201791Z 2025-09-07T08:07:07.2201890Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2202248Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2202570Z return mod(**inputs) 2025-09-07T08:07:07.2202919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2203474Z outputs = self.bert( 2025-09-07T08:07:07.2203827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2204211Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2204699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2205095Z layer_outputs = layer_module( 2025-09-07T08:07:07.2205438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2205797Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2206178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2206560Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2206934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2207304Z return func(*args, **kwargs) 2025-09-07T08:07:07.2207664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2208038Z self_outputs = self.self( 2025-09-07T08:07:07.2208391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2208753Z return func(*args, **kwargs) 2025-09-07T08:07:07.2209111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2209486Z self.key(current_states) 2025-09-07T08:07:07.2209595Z 2025-09-07T08:07:07.2209695Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2210052Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2210368Z return mod(**inputs) 2025-09-07T08:07:07.2210718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2211091Z outputs = self.bert( 2025-09-07T08:07:07.2211433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2211816Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2212189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2212568Z layer_outputs = layer_module( 2025-09-07T08:07:07.2212905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2213262Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2213643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2214028Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2214400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2214763Z return func(*args, **kwargs) 2025-09-07T08:07:07.2215121Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2215643Z self_outputs = self.self( 2025-09-07T08:07:07.2215997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2216366Z return func(*args, **kwargs) 2025-09-07T08:07:07.2216716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2217089Z self.value(current_states) 2025-09-07T08:07:07.2217200Z 2025-09-07T08:07:07.2217284Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2217563Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2217913Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2218229Z return mod(**inputs) 2025-09-07T08:07:07.2218646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2219026Z outputs = self.bert( 2025-09-07T08:07:07.2219371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2219744Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2220110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2220487Z layer_outputs = layer_module( 2025-09-07T08:07:07.2220824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2221190Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2221567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2221949Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2222333Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2222702Z return func(*args, **kwargs) 2025-09-07T08:07:07.2223048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2223418Z self_outputs = self.self( 2025-09-07T08:07:07.2223764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2224130Z return func(*args, **kwargs) 2025-09-07T08:07:07.2224478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2224912Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2225101Z 2025-09-07T08:07:07.2225198Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2225547Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2225866Z return mod(**inputs) 2025-09-07T08:07:07.2226206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2226580Z outputs = self.bert( 2025-09-07T08:07:07.2226926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2227309Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2227675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2228041Z layer_outputs = layer_module( 2025-09-07T08:07:07.2228379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2228754Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2229213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2229591Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2229974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2230343Z return func(*args, **kwargs) 2025-09-07T08:07:07.2230694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2231120Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2231533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2231911Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2232042Z 2025-09-07T08:07:07.2232138Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2232560Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2232871Z return mod(**inputs) 2025-09-07T08:07:07.2233209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2233574Z outputs = self.bert( 2025-09-07T08:07:07.2233906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2234287Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2234643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2235004Z layer_outputs = layer_module( 2025-09-07T08:07:07.2235325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2235662Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2236027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2236402Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2236792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2237177Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2237572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2238019Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2238441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2238834Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2238963Z 2025-09-07T08:07:07.2239061Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2239412Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2239716Z return mod(**inputs) 2025-09-07T08:07:07.2240060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2240427Z outputs = self.bert( 2025-09-07T08:07:07.2240758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2241126Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2241484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2241848Z layer_outputs = layer_module( 2025-09-07T08:07:07.2242175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2242518Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2242958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2243331Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2243717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2244097Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2244492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2244932Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2245340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2245744Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2246178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2246514Z return self.act(input) 2025-09-07T08:07:07.2246619Z 2025-09-07T08:07:07.2246713Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2247063Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2247378Z return mod(**inputs) 2025-09-07T08:07:07.2247729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2248101Z outputs = self.bert( 2025-09-07T08:07:07.2248434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2248806Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2249158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2249525Z layer_outputs = layer_module( 2025-09-07T08:07:07.2249848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2250190Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2250554Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2250924Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2251309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2251688Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2252078Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2252529Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2252958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2253330Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2253459Z 2025-09-07T08:07:07.2253550Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2253888Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2254194Z return mod(**inputs) 2025-09-07T08:07:07.2254525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2254887Z outputs = self.bert( 2025-09-07T08:07:07.2255218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2255588Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2255942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2256402Z layer_outputs = layer_module( 2025-09-07T08:07:07.2256732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2257078Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2257438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2257868Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2258233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2258592Z return func(*args, **kwargs) 2025-09-07T08:07:07.2258941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2259302Z self_outputs = self.self( 2025-09-07T08:07:07.2259707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2260074Z return func(*args, **kwargs) 2025-09-07T08:07:07.2260419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2260925Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2261185Z 2025-09-07T08:07:07.2261276Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2261615Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2261921Z return mod(**inputs) 2025-09-07T08:07:07.2262256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2262617Z outputs = self.bert( 2025-09-07T08:07:07.2262947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2263316Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2263667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2264035Z layer_outputs = layer_module( 2025-09-07T08:07:07.2264363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2264705Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2265065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2265440Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2265803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2266159Z return func(*args, **kwargs) 2025-09-07T08:07:07.2266507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2266868Z self_outputs = self.self( 2025-09-07T08:07:07.2267204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2267557Z return func(*args, **kwargs) 2025-09-07T08:07:07.2267895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2268258Z self.key(current_states) 2025-09-07T08:07:07.2268367Z 2025-09-07T08:07:07.2268463Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2268808Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2269132Z return mod(**inputs) 2025-09-07T08:07:07.2269481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2269931Z outputs = self.bert( 2025-09-07T08:07:07.2270267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2270639Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2270997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2271363Z layer_outputs = layer_module( 2025-09-07T08:07:07.2271693Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2272040Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2272404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2272778Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2273426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2273795Z return func(*args, **kwargs) 2025-09-07T08:07:07.2274144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2274506Z self_outputs = self.self( 2025-09-07T08:07:07.2274847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2275199Z return func(*args, **kwargs) 2025-09-07T08:07:07.2275543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2275908Z self.value(current_states) 2025-09-07T08:07:07.2276014Z 2025-09-07T08:07:07.2276088Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2276306Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2276652Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2276959Z return mod(**inputs) 2025-09-07T08:07:07.2277297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2277663Z outputs = self.bert( 2025-09-07T08:07:07.2277995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2278362Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2278716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2279082Z layer_outputs = layer_module( 2025-09-07T08:07:07.2279409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2279756Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2280125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2280493Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2280859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2281218Z return func(*args, **kwargs) 2025-09-07T08:07:07.2281571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2281933Z self_outputs = self.self( 2025-09-07T08:07:07.2282274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2282634Z return func(*args, **kwargs) 2025-09-07T08:07:07.2282977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2283498Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2283680Z 2025-09-07T08:07:07.2283773Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2284109Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2284412Z return mod(**inputs) 2025-09-07T08:07:07.2284748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2285111Z outputs = self.bert( 2025-09-07T08:07:07.2285442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2285808Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2286158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2286520Z layer_outputs = layer_module( 2025-09-07T08:07:07.2286917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2287265Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2287629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2288002Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2288359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2288719Z return func(*args, **kwargs) 2025-09-07T08:07:07.2289071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2289487Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2289901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2290297Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2290431Z 2025-09-07T08:07:07.2290526Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2290862Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2291168Z return mod(**inputs) 2025-09-07T08:07:07.2291502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2291864Z outputs = self.bert( 2025-09-07T08:07:07.2292197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2292564Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2292918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2293282Z layer_outputs = layer_module( 2025-09-07T08:07:07.2293608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2293948Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2294307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2294685Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2295075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2295456Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2295852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2296295Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2296711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2297224Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2297383Z 2025-09-07T08:07:07.2297509Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2297924Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2298246Z return mod(**inputs) 2025-09-07T08:07:07.2298586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2298950Z outputs = self.bert( 2025-09-07T08:07:07.2299304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2299671Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2300029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2300462Z layer_outputs = layer_module( 2025-09-07T08:07:07.2300785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2301146Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2301524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2301897Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2302279Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2302669Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2303200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2303664Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2304080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2304487Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2304862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2305197Z return self.act(input) 2025-09-07T08:07:07.2305301Z 2025-09-07T08:07:07.2305396Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2305751Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2306056Z return mod(**inputs) 2025-09-07T08:07:07.2306413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2306795Z outputs = self.bert( 2025-09-07T08:07:07.2307134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2307532Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2307883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2308256Z layer_outputs = layer_module( 2025-09-07T08:07:07.2308584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2308928Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2309290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2309662Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2310053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2315013Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2315443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2315996Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2316430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2316831Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2316967Z 2025-09-07T08:07:07.2317085Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2317440Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2317766Z return mod(**inputs) 2025-09-07T08:07:07.2318125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2318496Z outputs = self.bert( 2025-09-07T08:07:07.2318953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2319349Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2319722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2320099Z layer_outputs = layer_module( 2025-09-07T08:07:07.2320435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2320803Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2321205Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2321585Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2321984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2322353Z return func(*args, **kwargs) 2025-09-07T08:07:07.2322726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2323103Z self_outputs = self.self( 2025-09-07T08:07:07.2323454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2323819Z return func(*args, **kwargs) 2025-09-07T08:07:07.2324185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2324712Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2324975Z 2025-09-07T08:07:07.2325075Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2325454Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2325785Z return mod(**inputs) 2025-09-07T08:07:07.2326138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2326519Z outputs = self.bert( 2025-09-07T08:07:07.2326868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2327242Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2327612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2327982Z layer_outputs = layer_module( 2025-09-07T08:07:07.2328314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2328672Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2329058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2329525Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2329946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2330323Z return func(*args, **kwargs) 2025-09-07T08:07:07.2330675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2331059Z self_outputs = self.self( 2025-09-07T08:07:07.2331404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2331760Z return func(*args, **kwargs) 2025-09-07T08:07:07.2332137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2332504Z self.key(current_states) 2025-09-07T08:07:07.2332611Z 2025-09-07T08:07:07.2332735Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2333167Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2333484Z return mod(**inputs) 2025-09-07T08:07:07.2333831Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2334216Z outputs = self.bert( 2025-09-07T08:07:07.2334570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2334960Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2335323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2335689Z layer_outputs = layer_module( 2025-09-07T08:07:07.2336034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2336395Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2336793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2337190Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2337607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2338007Z return func(*args, **kwargs) 2025-09-07T08:07:07.2338370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2338752Z self_outputs = self.self( 2025-09-07T08:07:07.2339099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2339470Z return func(*args, **kwargs) 2025-09-07T08:07:07.2339815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2340190Z self.value(current_states) 2025-09-07T08:07:07.2340299Z 2025-09-07T08:07:07.2340377Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2340595Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2340932Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2341236Z return mod(**inputs) 2025-09-07T08:07:07.2341575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2341936Z outputs = self.bert( 2025-09-07T08:07:07.2342268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2342635Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2342988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2343472Z layer_outputs = layer_module( 2025-09-07T08:07:07.2343849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2344199Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2344569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2344943Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2345309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2345666Z return func(*args, **kwargs) 2025-09-07T08:07:07.2346015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2346379Z self_outputs = self.self( 2025-09-07T08:07:07.2346722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2347144Z return func(*args, **kwargs) 2025-09-07T08:07:07.2347488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2347911Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2348092Z 2025-09-07T08:07:07.2348186Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2348527Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2348831Z return mod(**inputs) 2025-09-07T08:07:07.2349168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2349532Z outputs = self.bert( 2025-09-07T08:07:07.2349862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2350231Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2350586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2350950Z layer_outputs = layer_module( 2025-09-07T08:07:07.2351277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2351623Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2351985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2352355Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2352722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2353087Z return func(*args, **kwargs) 2025-09-07T08:07:07.2353444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2353875Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2354289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2354676Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2354815Z 2025-09-07T08:07:07.2354913Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2355261Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2355567Z return mod(**inputs) 2025-09-07T08:07:07.2355911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2356278Z outputs = self.bert( 2025-09-07T08:07:07.2356617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2357087Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2357461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2357840Z layer_outputs = layer_module( 2025-09-07T08:07:07.2358181Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2358535Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2358907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2359288Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2359685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2360080Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2360568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2361028Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2361450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2361835Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2361973Z 2025-09-07T08:07:07.2362074Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2362421Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2362731Z return mod(**inputs) 2025-09-07T08:07:07.2363091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2363459Z outputs = self.bert( 2025-09-07T08:07:07.2363797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2364174Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2364536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2364906Z layer_outputs = layer_module( 2025-09-07T08:07:07.2365240Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2365589Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2365952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2366329Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2366717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2367104Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2367508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2367952Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2368385Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2368802Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2369182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2369517Z return self.act(input) 2025-09-07T08:07:07.2369629Z 2025-09-07T08:07:07.2369726Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2370080Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2370400Z return mod(**inputs) 2025-09-07T08:07:07.2370806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2371210Z outputs = self.bert( 2025-09-07T08:07:07.2371565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2371945Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2372316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2372691Z layer_outputs = layer_module( 2025-09-07T08:07:07.2373024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2373381Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2373758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2374146Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2374602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2375002Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2375407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2375871Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2376300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2376676Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2376812Z 2025-09-07T08:07:07.2376906Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2377256Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2377604Z return mod(**inputs) 2025-09-07T08:07:07.2377946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2378325Z outputs = self.bert( 2025-09-07T08:07:07.2378663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2379032Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2379387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2379749Z layer_outputs = layer_module( 2025-09-07T08:07:07.2380077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2380429Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2380803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2381186Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2381559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2381635Z return func(*args, **kwargs) 2025-09-07T08:07:07.2381876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2381943Z self_outputs = self.self( 2025-09-07T08:07:07.2382180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2382240Z return func(*args, **kwargs) 2025-09-07T08:07:07.2382483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2382690Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2382765Z 2025-09-07T08:07:07.2382869Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2383105Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2383174Z return mod(**inputs) 2025-09-07T08:07:07.2383423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2383492Z outputs = self.bert( 2025-09-07T08:07:07.2383747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2383821Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2384070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2384131Z layer_outputs = layer_module( 2025-09-07T08:07:07.2384343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2384417Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2384746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2384820Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2385058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2385118Z return func(*args, **kwargs) 2025-09-07T08:07:07.2385362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2385426Z self_outputs = self.self( 2025-09-07T08:07:07.2385663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2385721Z return func(*args, **kwargs) 2025-09-07T08:07:07.2385967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2386032Z self.key(current_states) 2025-09-07T08:07:07.2386036Z 2025-09-07T08:07:07.2386125Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2386312Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2386368Z return mod(**inputs) 2025-09-07T08:07:07.2386617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2386676Z outputs = self.bert( 2025-09-07T08:07:07.2386920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2386990Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2387232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2387306Z layer_outputs = layer_module( 2025-09-07T08:07:07.2387522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2387601Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2387874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2387944Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2388218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2388284Z return func(*args, **kwargs) 2025-09-07T08:07:07.2388529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2388590Z self_outputs = self.self( 2025-09-07T08:07:07.2388825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2388973Z return func(*args, **kwargs) 2025-09-07T08:07:07.2389215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2389282Z self.value(current_states) 2025-09-07T08:07:07.2389285Z 2025-09-07T08:07:07.2389357Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2389447Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2389633Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2389691Z return mod(**inputs) 2025-09-07T08:07:07.2389933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2389992Z outputs = self.bert( 2025-09-07T08:07:07.2390234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2390303Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2390606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2390690Z layer_outputs = layer_module( 2025-09-07T08:07:07.2390920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2390990Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2391252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2391321Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2391585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2391662Z return func(*args, **kwargs) 2025-09-07T08:07:07.2391916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2391995Z self_outputs = self.self( 2025-09-07T08:07:07.2392231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2392307Z return func(*args, **kwargs) 2025-09-07T08:07:07.2392548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2392693Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2392696Z 2025-09-07T08:07:07.2392784Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2392983Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2393060Z return mod(**inputs) 2025-09-07T08:07:07.2393308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2393369Z outputs = self.bert( 2025-09-07T08:07:07.2393612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2393674Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2393915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2393975Z layer_outputs = layer_module( 2025-09-07T08:07:07.2394204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2394276Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2394518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2394604Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2394902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2394999Z return func(*args, **kwargs) 2025-09-07T08:07:07.2395241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2395384Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2395624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2395699Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2395705Z 2025-09-07T08:07:07.2395797Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2395991Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2396058Z return mod(**inputs) 2025-09-07T08:07:07.2396313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2396442Z outputs = self.bert( 2025-09-07T08:07:07.2396693Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2396758Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2397013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2397077Z layer_outputs = layer_module( 2025-09-07T08:07:07.2397294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2397365Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2397606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2397702Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2397971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2398047Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2398326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2398442Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2398689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2398760Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2398764Z 2025-09-07T08:07:07.2398881Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2399072Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2399134Z return mod(**inputs) 2025-09-07T08:07:07.2399383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2399444Z outputs = self.bert( 2025-09-07T08:07:07.2399694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2399774Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2400024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2400087Z layer_outputs = layer_module( 2025-09-07T08:07:07.2400302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2400379Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2400622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2400704Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2401012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2401121Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2401398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2401508Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2401756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2401858Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2402091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2402156Z return self.act(input) 2025-09-07T08:07:07.2402159Z 2025-09-07T08:07:07.2402251Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2402545Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2402604Z return mod(**inputs) 2025-09-07T08:07:07.2402852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2402924Z outputs = self.bert( 2025-09-07T08:07:07.2403314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2403381Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2403621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2403686Z layer_outputs = layer_module( 2025-09-07T08:07:07.2403910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2403990Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2404236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2404310Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2404574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2404642Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2404922Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2405047Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2405288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2405366Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2405372Z 2025-09-07T08:07:07.2405463Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2405656Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2405712Z return mod(**inputs) 2025-09-07T08:07:07.2405975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2406030Z outputs = self.bert( 2025-09-07T08:07:07.2406271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2406335Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2406573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2406641Z layer_outputs = layer_module( 2025-09-07T08:07:07.2406853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2407014Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2407329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2407413Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2407662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2407725Z return func(*args, **kwargs) 2025-09-07T08:07:07.2407968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2408052Z self_outputs = self.self( 2025-09-07T08:07:07.2408291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2408353Z return func(*args, **kwargs) 2025-09-07T08:07:07.2408594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2408916Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2408919Z 2025-09-07T08:07:07.2409017Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2409210Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2409278Z return mod(**inputs) 2025-09-07T08:07:07.2409528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2409593Z outputs = self.bert( 2025-09-07T08:07:07.2409839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2409907Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2410159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2410231Z layer_outputs = layer_module( 2025-09-07T08:07:07.2410457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2410527Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2410775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2410847Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2411088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2411165Z return func(*args, **kwargs) 2025-09-07T08:07:07.2411406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2411468Z self_outputs = self.self( 2025-09-07T08:07:07.2411708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2411771Z return func(*args, **kwargs) 2025-09-07T08:07:07.2412035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2412098Z self.key(current_states) 2025-09-07T08:07:07.2412101Z 2025-09-07T08:07:07.2412193Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2412376Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2412431Z return mod(**inputs) 2025-09-07T08:07:07.2412695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2412753Z outputs = self.bert( 2025-09-07T08:07:07.2412996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2413142Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2413386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2413453Z layer_outputs = layer_module( 2025-09-07T08:07:07.2413667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2413744Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2413986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2414065Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2414301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2414363Z return func(*args, **kwargs) 2025-09-07T08:07:07.2414612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2414742Z self_outputs = self.self( 2025-09-07T08:07:07.2414988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2415048Z return func(*args, **kwargs) 2025-09-07T08:07:07.2415291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2415363Z self.value(current_states) 2025-09-07T08:07:07.2415366Z 2025-09-07T08:07:07.2415439Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2415537Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2415718Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2415776Z return mod(**inputs) 2025-09-07T08:07:07.2416031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2416097Z outputs = self.bert( 2025-09-07T08:07:07.2416349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2416415Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2416661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2416723Z layer_outputs = layer_module( 2025-09-07T08:07:07.2416937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2417013Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2417252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2417329Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2417619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2417684Z return func(*args, **kwargs) 2025-09-07T08:07:07.2417934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2417995Z self_outputs = self.self( 2025-09-07T08:07:07.2418241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2418300Z return func(*args, **kwargs) 2025-09-07T08:07:07.2418543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2418671Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2418674Z 2025-09-07T08:07:07.2418764Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2419004Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2419100Z return mod(**inputs) 2025-09-07T08:07:07.2419357Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2419416Z outputs = self.bert( 2025-09-07T08:07:07.2419662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2419735Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2419977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2420046Z layer_outputs = layer_module( 2025-09-07T08:07:07.2420259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2420327Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2420640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2420714Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2420960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2421025Z return func(*args, **kwargs) 2025-09-07T08:07:07.2421266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2421394Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2421640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2421724Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2421727Z 2025-09-07T08:07:07.2421817Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2422011Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2422069Z return mod(**inputs) 2025-09-07T08:07:07.2422312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2422400Z outputs = self.bert( 2025-09-07T08:07:07.2422644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2422721Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2422964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2423027Z layer_outputs = layer_module( 2025-09-07T08:07:07.2423247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2423321Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2423571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2423647Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2423916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2423988Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2424265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2424382Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2424628Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2424711Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2424714Z 2025-09-07T08:07:07.2424808Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2425040Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2425139Z return mod(**inputs) 2025-09-07T08:07:07.2425387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2425456Z outputs = self.bert( 2025-09-07T08:07:07.2425703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2425768Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2426019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2426082Z layer_outputs = layer_module( 2025-09-07T08:07:07.2426304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2426374Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2426681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2426756Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2427015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2427089Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2427362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2427489Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2427735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2427834Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2428054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2428122Z return self.act(input) 2025-09-07T08:07:07.2428125Z 2025-09-07T08:07:07.2428224Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2428471Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2428606Z return mod(**inputs) 2025-09-07T08:07:07.2428886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2428976Z outputs = self.bert( 2025-09-07T08:07:07.2429282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2429377Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2429813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2429948Z layer_outputs = layer_module( 2025-09-07T08:07:07.2430197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2430295Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2430596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2430686Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2431074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2431170Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2431505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2431657Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2432028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2432219Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2432223Z 2025-09-07T08:07:07.2432355Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2432602Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2432688Z return mod(**inputs) 2025-09-07T08:07:07.2432979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2433083Z outputs = self.bert( 2025-09-07T08:07:07.2433397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2433534Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2433806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2433944Z layer_outputs = layer_module( 2025-09-07T08:07:07.2434249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2434335Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2434695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2434796Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2435110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2435203Z return func(*args, **kwargs) 2025-09-07T08:07:07.2435474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2435630Z self_outputs = self.self( 2025-09-07T08:07:07.2435917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2436057Z return func(*args, **kwargs) 2025-09-07T08:07:07.2436352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2436612Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2436616Z 2025-09-07T08:07:07.2436723Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2436985Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2437127Z return mod(**inputs) 2025-09-07T08:07:07.2437406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2437522Z outputs = self.bert( 2025-09-07T08:07:07.2437797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2437970Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2438260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2438367Z layer_outputs = layer_module( 2025-09-07T08:07:07.2438647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2438745Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2439034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2439176Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2439472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2439594Z return func(*args, **kwargs) 2025-09-07T08:07:07.2439911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2440064Z self_outputs = self.self( 2025-09-07T08:07:07.2440323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2440507Z return func(*args, **kwargs) 2025-09-07T08:07:07.2440787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2440908Z self.key(current_states) 2025-09-07T08:07:07.2440911Z 2025-09-07T08:07:07.2441066Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2441283Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2441443Z return mod(**inputs) 2025-09-07T08:07:07.2441731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2441881Z outputs = self.bert( 2025-09-07T08:07:07.2442200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2442296Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2442599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2442733Z layer_outputs = layer_module( 2025-09-07T08:07:07.2450623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2450717Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2451004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2451086Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2451371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2451446Z return func(*args, **kwargs) 2025-09-07T08:07:07.2451701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2451776Z self_outputs = self.self( 2025-09-07T08:07:07.2452018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2452086Z return func(*args, **kwargs) 2025-09-07T08:07:07.2452342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2452407Z self.value(current_states) 2025-09-07T08:07:07.2452412Z 2025-09-07T08:07:07.2452494Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2452596Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2452808Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2452874Z return mod(**inputs) 2025-09-07T08:07:07.2453145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2453211Z outputs = self.bert( 2025-09-07T08:07:07.2453460Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2453536Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2453779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2453842Z layer_outputs = layer_module( 2025-09-07T08:07:07.2454066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2454144Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2454515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2454633Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2454883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2454947Z return func(*args, **kwargs) 2025-09-07T08:07:07.2455189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2455258Z self_outputs = self.self( 2025-09-07T08:07:07.2455494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2455555Z return func(*args, **kwargs) 2025-09-07T08:07:07.2455792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2455921Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2455927Z 2025-09-07T08:07:07.2456092Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2456291Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2456352Z return mod(**inputs) 2025-09-07T08:07:07.2456600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2456660Z outputs = self.bert( 2025-09-07T08:07:07.2456903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2456971Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2457216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2457280Z layer_outputs = layer_module( 2025-09-07T08:07:07.2457508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2457643Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2457884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2457957Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2458194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2458261Z return func(*args, **kwargs) 2025-09-07T08:07:07.2458500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2458619Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2458860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2458937Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2458945Z 2025-09-07T08:07:07.2459043Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2459232Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2459292Z return mod(**inputs) 2025-09-07T08:07:07.2459535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2459594Z outputs = self.bert( 2025-09-07T08:07:07.2459843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2459911Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2460156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2460216Z layer_outputs = layer_module( 2025-09-07T08:07:07.2460488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2460599Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2460841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2460916Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2461179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2461248Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2461542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2461658Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2461908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2462048Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2462053Z 2025-09-07T08:07:07.2462150Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2462340Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2462396Z return mod(**inputs) 2025-09-07T08:07:07.2462648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2462707Z outputs = self.bert( 2025-09-07T08:07:07.2462960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2463025Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2463266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2463337Z layer_outputs = layer_module( 2025-09-07T08:07:07.2463555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2463635Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2463877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2463956Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2464219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2464287Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2464566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2464673Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2464923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2465032Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2465242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2465312Z return self.act(input) 2025-09-07T08:07:07.2465316Z 2025-09-07T08:07:07.2465414Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2465609Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2465667Z return mod(**inputs) 2025-09-07T08:07:07.2465913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2465975Z outputs = self.bert( 2025-09-07T08:07:07.2466219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2466353Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2466635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2466705Z layer_outputs = layer_module( 2025-09-07T08:07:07.2466925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2466999Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2467254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2467331Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2467594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2467662Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2467942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2468153Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2468406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2468490Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2468494Z 2025-09-07T08:07:07.2468593Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2468791Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2468849Z return mod(**inputs) 2025-09-07T08:07:07.2469099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2469163Z outputs = self.bert( 2025-09-07T08:07:07.2469408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2469481Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2469722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2469785Z layer_outputs = layer_module( 2025-09-07T08:07:07.2470005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2470074Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2470322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2470393Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2470639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2470704Z return func(*args, **kwargs) 2025-09-07T08:07:07.2470956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2471024Z self_outputs = self.self( 2025-09-07T08:07:07.2471263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2471328Z return func(*args, **kwargs) 2025-09-07T08:07:07.2471570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2471778Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2471781Z 2025-09-07T08:07:07.2471881Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2472075Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2472134Z return mod(**inputs) 2025-09-07T08:07:07.2472424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2472513Z outputs = self.bert( 2025-09-07T08:07:07.2472762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2472830Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2473072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2473140Z layer_outputs = layer_module( 2025-09-07T08:07:07.2473363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2473436Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2473677Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2473755Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2474052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2474122Z return func(*args, **kwargs) 2025-09-07T08:07:07.2474362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2474424Z self_outputs = self.self( 2025-09-07T08:07:07.2474668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2474729Z return func(*args, **kwargs) 2025-09-07T08:07:07.2474974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2475035Z self.key(current_states) 2025-09-07T08:07:07.2475039Z 2025-09-07T08:07:07.2475130Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2475328Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2475390Z return mod(**inputs) 2025-09-07T08:07:07.2475638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2475695Z outputs = self.bert( 2025-09-07T08:07:07.2475945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2476010Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2476249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2476316Z layer_outputs = layer_module( 2025-09-07T08:07:07.2476527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2476596Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2476840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2476913Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2477153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2477211Z return func(*args, **kwargs) 2025-09-07T08:07:07.2477455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2477516Z self_outputs = self.self( 2025-09-07T08:07:07.2477750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2477810Z return func(*args, **kwargs) 2025-09-07T08:07:07.2478050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2478154Z self.value(current_states) 2025-09-07T08:07:07.2478186Z 2025-09-07T08:07:07.2478262Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2478355Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2478540Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2478599Z return mod(**inputs) 2025-09-07T08:07:07.2478851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2478907Z outputs = self.bert( 2025-09-07T08:07:07.2479154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2479218Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2479458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2479525Z layer_outputs = layer_module( 2025-09-07T08:07:07.2479796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2479868Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2480111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2480182Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2480421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2480481Z return func(*args, **kwargs) 2025-09-07T08:07:07.2480720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2480780Z self_outputs = self.self( 2025-09-07T08:07:07.2481013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2481077Z return func(*args, **kwargs) 2025-09-07T08:07:07.2481319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2481447Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2481450Z 2025-09-07T08:07:07.2481539Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2481729Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2481783Z return mod(**inputs) 2025-09-07T08:07:07.2482026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2482086Z outputs = self.bert( 2025-09-07T08:07:07.2482329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2482395Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2482637Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2482699Z layer_outputs = layer_module( 2025-09-07T08:07:07.2482912Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2482976Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2483216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2483288Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2483519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2483579Z return func(*args, **kwargs) 2025-09-07T08:07:07.2483817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2483987Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2484266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2484342Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2484346Z 2025-09-07T08:07:07.2484434Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2484618Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2484676Z return mod(**inputs) 2025-09-07T08:07:07.2484919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2484976Z outputs = self.bert( 2025-09-07T08:07:07.2485218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2485281Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2485590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2485652Z layer_outputs = layer_module( 2025-09-07T08:07:07.2485867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2485933Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2486174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2486247Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2486507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2486577Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2486852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2486972Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2487212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2487283Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2487287Z 2025-09-07T08:07:07.2487377Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2487561Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2487619Z return mod(**inputs) 2025-09-07T08:07:07.2487860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2487920Z outputs = self.bert( 2025-09-07T08:07:07.2488166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2488229Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2488474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2488534Z layer_outputs = layer_module( 2025-09-07T08:07:07.2488748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2488813Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2489052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2489128Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2489385Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2489455Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2489731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2489941Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2490184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2490287Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2490499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2490561Z return self.act(input) 2025-09-07T08:07:07.2490565Z 2025-09-07T08:07:07.2490656Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2490842Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2490899Z return mod(**inputs) 2025-09-07T08:07:07.2491146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2491205Z outputs = self.bert( 2025-09-07T08:07:07.2491517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2491581Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2491824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2491892Z layer_outputs = layer_module( 2025-09-07T08:07:07.2492104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2492175Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2492414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2492487Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2492749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2492815Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2493093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2493218Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2493467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2493541Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2493544Z 2025-09-07T08:07:07.2493636Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2493825Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2493882Z return mod(**inputs) 2025-09-07T08:07:07.2494128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2494192Z outputs = self.bert( 2025-09-07T08:07:07.2494437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2494506Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2494747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2494814Z layer_outputs = layer_module( 2025-09-07T08:07:07.2495032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2495099Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2495339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2495409Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2495694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2495786Z return func(*args, **kwargs) 2025-09-07T08:07:07.2496029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2496092Z self_outputs = self.self( 2025-09-07T08:07:07.2496328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2496388Z return func(*args, **kwargs) 2025-09-07T08:07:07.2496628Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2496829Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2496833Z 2025-09-07T08:07:07.2496927Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2497173Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2497233Z return mod(**inputs) 2025-09-07T08:07:07.2497478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2497571Z outputs = self.bert( 2025-09-07T08:07:07.2497816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2497880Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2498121Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2498180Z layer_outputs = layer_module( 2025-09-07T08:07:07.2498397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2498464Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2498707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2498778Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2499011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2499072Z return func(*args, **kwargs) 2025-09-07T08:07:07.2499309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2499369Z self_outputs = self.self( 2025-09-07T08:07:07.2499604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2499663Z return func(*args, **kwargs) 2025-09-07T08:07:07.2499901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2499965Z self.key(current_states) 2025-09-07T08:07:07.2499971Z 2025-09-07T08:07:07.2500061Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2500245Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2500300Z return mod(**inputs) 2025-09-07T08:07:07.2500550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2500605Z outputs = self.bert( 2025-09-07T08:07:07.2500849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2500912Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2501150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2501218Z layer_outputs = layer_module( 2025-09-07T08:07:07.2501479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2501980Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2502222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2502292Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2502526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2502586Z return func(*args, **kwargs) 2025-09-07T08:07:07.2502825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2502888Z self_outputs = self.self( 2025-09-07T08:07:07.2503316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2503384Z return func(*args, **kwargs) 2025-09-07T08:07:07.2503805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2503877Z self.value(current_states) 2025-09-07T08:07:07.2503881Z 2025-09-07T08:07:07.2503955Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2504052Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2504240Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2504295Z return mod(**inputs) 2025-09-07T08:07:07.2504540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2504596Z outputs = self.bert( 2025-09-07T08:07:07.2504841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2504907Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2505153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2505220Z layer_outputs = layer_module( 2025-09-07T08:07:07.2505434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2505502Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2505741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2505811Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2506047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2506106Z return func(*args, **kwargs) 2025-09-07T08:07:07.2506349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2506411Z self_outputs = self.self( 2025-09-07T08:07:07.2506651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2506711Z return func(*args, **kwargs) 2025-09-07T08:07:07.2506952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2507079Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2507082Z 2025-09-07T08:07:07.2507175Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2507364Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2507420Z return mod(**inputs) 2025-09-07T08:07:07.2507662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2507773Z outputs = self.bert( 2025-09-07T08:07:07.2508081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2508153Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2508400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2508465Z layer_outputs = layer_module( 2025-09-07T08:07:07.2508685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2508759Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2509004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2509075Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2509323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2509389Z return func(*args, **kwargs) 2025-09-07T08:07:07.2509688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2509816Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2510060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2510142Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2510146Z 2025-09-07T08:07:07.2510243Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2510438Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2510503Z return mod(**inputs) 2025-09-07T08:07:07.2510754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2510815Z outputs = self.bert( 2025-09-07T08:07:07.2511062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2511133Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2511372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2511435Z layer_outputs = layer_module( 2025-09-07T08:07:07.2511653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2511724Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2511964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2512038Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2512298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2512375Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2512652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2512770Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2513012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2513084Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2513089Z 2025-09-07T08:07:07.2513184Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2513371Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2513431Z return mod(**inputs) 2025-09-07T08:07:07.2513673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2513809Z outputs = self.bert( 2025-09-07T08:07:07.2514053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2514119Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2514362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2514425Z layer_outputs = layer_module( 2025-09-07T08:07:07.2514643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2514712Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2514951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2515024Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2515283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2515417Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2515695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2515804Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2516045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2516147Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2516359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2516421Z return self.act(input) 2025-09-07T08:07:07.2516425Z 2025-09-07T08:07:07.2516518Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2516705Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2516765Z return mod(**inputs) 2025-09-07T08:07:07.2517010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2517065Z outputs = self.bert( 2025-09-07T08:07:07.2517314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2517376Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2517616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2517685Z layer_outputs = layer_module( 2025-09-07T08:07:07.2517898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2517969Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2518212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2518287Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2518543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2518609Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2518886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2519012Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2519259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2519330Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2519333Z 2025-09-07T08:07:07.2519424Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2519653Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2519741Z return mod(**inputs) 2025-09-07T08:07:07.2519998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2520054Z outputs = self.bert( 2025-09-07T08:07:07.2520298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2520361Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2520600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2520663Z layer_outputs = layer_module( 2025-09-07T08:07:07.2520875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2520946Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2521243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2521320Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2521568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2521631Z return func(*args, **kwargs) 2025-09-07T08:07:07.2521876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2521939Z self_outputs = self.self( 2025-09-07T08:07:07.2522176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2522237Z return func(*args, **kwargs) 2025-09-07T08:07:07.2522478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2522686Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2522690Z 2025-09-07T08:07:07.2522779Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2522965Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2523021Z return mod(**inputs) 2025-09-07T08:07:07.2523270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2523330Z outputs = self.bert( 2025-09-07T08:07:07.2523573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2523636Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2523874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2523937Z layer_outputs = layer_module( 2025-09-07T08:07:07.2524156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2524224Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2524465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2524535Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2524772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2524840Z return func(*args, **kwargs) 2025-09-07T08:07:07.2525077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2525141Z self_outputs = self.self( 2025-09-07T08:07:07.2525374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2525500Z return func(*args, **kwargs) 2025-09-07T08:07:07.2525739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2525800Z self.key(current_states) 2025-09-07T08:07:07.2525803Z 2025-09-07T08:07:07.2525896Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2526080Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2526141Z return mod(**inputs) 2025-09-07T08:07:07.2526391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2526449Z outputs = self.bert( 2025-09-07T08:07:07.2526696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2526760Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2527070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2527133Z layer_outputs = layer_module( 2025-09-07T08:07:07.2527349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2527418Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2527663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2527735Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2527969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2528033Z return func(*args, **kwargs) 2025-09-07T08:07:07.2528272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2528337Z self_outputs = self.self( 2025-09-07T08:07:07.2528579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2528640Z return func(*args, **kwargs) 2025-09-07T08:07:07.2528885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2528947Z self.value(current_states) 2025-09-07T08:07:07.2528950Z 2025-09-07T08:07:07.2529025Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2529117Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2529299Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2529361Z return mod(**inputs) 2025-09-07T08:07:07.2529605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2529665Z outputs = self.bert( 2025-09-07T08:07:07.2529919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2529983Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2530227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2530286Z layer_outputs = layer_module( 2025-09-07T08:07:07.2530504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2530569Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2530807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2530880Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2531115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2531239Z return func(*args, **kwargs) 2025-09-07T08:07:07.2531479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2531538Z self_outputs = self.self( 2025-09-07T08:07:07.2531773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2531831Z return func(*args, **kwargs) 2025-09-07T08:07:07.2532072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2532194Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2532198Z 2025-09-07T08:07:07.2532285Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2532467Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2532524Z return mod(**inputs) 2025-09-07T08:07:07.2532843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2532900Z outputs = self.bert( 2025-09-07T08:07:07.2533145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2533209Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2533448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2533515Z layer_outputs = layer_module( 2025-09-07T08:07:07.2533726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2533794Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2534033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2534110Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2534347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2534405Z return func(*args, **kwargs) 2025-09-07T08:07:07.2534647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2534765Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2535005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2535081Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2535084Z 2025-09-07T08:07:07.2535171Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2535356Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2535413Z return mod(**inputs) 2025-09-07T08:07:07.2535663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2535721Z outputs = self.bert( 2025-09-07T08:07:07.2535962Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2536026Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2536267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2536334Z layer_outputs = layer_module( 2025-09-07T08:07:07.2536547Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2536614Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2536861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2537002Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2537288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2537353Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2537697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2537806Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2538046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2538117Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2538121Z 2025-09-07T08:07:07.2538209Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2538397Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2538544Z return mod(**inputs) 2025-09-07T08:07:07.2538794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2538856Z outputs = self.bert( 2025-09-07T08:07:07.2539099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2539163Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2539404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2539464Z layer_outputs = layer_module( 2025-09-07T08:07:07.2539676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2539742Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2539989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2540061Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2540323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2540389Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2540660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2540768Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2541008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2541113Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2541318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2541381Z return self.act(input) 2025-09-07T08:07:07.2541389Z 2025-09-07T08:07:07.2541479Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2541663Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2541722Z return mod(**inputs) 2025-09-07T08:07:07.2541965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2542022Z outputs = self.bert( 2025-09-07T08:07:07.2542264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2542327Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2542569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2542631Z layer_outputs = layer_module( 2025-09-07T08:07:07.2542901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2543002Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2543242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2543315Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2543569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2543636Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2543906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2544025Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2544269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2544411Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2544415Z 2025-09-07T08:07:07.2544514Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2544696Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2544760Z return mod(**inputs) 2025-09-07T08:07:07.2545005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2545061Z outputs = self.bert( 2025-09-07T08:07:07.2545308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2545372Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2545617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2545678Z layer_outputs = layer_module( 2025-09-07T08:07:07.2545894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2545965Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2546206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2546282Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2546517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2546585Z return func(*args, **kwargs) 2025-09-07T08:07:07.2546826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2546889Z self_outputs = self.self( 2025-09-07T08:07:07.2547132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2547195Z return func(*args, **kwargs) 2025-09-07T08:07:07.2547443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2547642Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2547645Z 2025-09-07T08:07:07.2547732Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2547917Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2547972Z return mod(**inputs) 2025-09-07T08:07:07.2548223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2548278Z outputs = self.bert( 2025-09-07T08:07:07.2548521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2548632Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2548908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2548971Z layer_outputs = layer_module( 2025-09-07T08:07:07.2549181Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2549250Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2549488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2549557Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2549794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2549851Z return func(*args, **kwargs) 2025-09-07T08:07:07.2550092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2550230Z self_outputs = self.self( 2025-09-07T08:07:07.2550470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2550531Z return func(*args, **kwargs) 2025-09-07T08:07:07.2550771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2550836Z self.key(current_states) 2025-09-07T08:07:07.2550840Z 2025-09-07T08:07:07.2550930Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2551127Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2551186Z return mod(**inputs) 2025-09-07T08:07:07.2551431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2551498Z outputs = self.bert( 2025-09-07T08:07:07.2551745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2551817Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2552061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2552125Z layer_outputs = layer_module( 2025-09-07T08:07:07.2552340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2552407Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2552652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2552721Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2552952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2553026Z return func(*args, **kwargs) 2025-09-07T08:07:07.2553266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2553333Z self_outputs = self.self( 2025-09-07T08:07:07.2553564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2553626Z return func(*args, **kwargs) 2025-09-07T08:07:07.2553872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2553937Z self.value(current_states) 2025-09-07T08:07:07.2553939Z 2025-09-07T08:07:07.2554017Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2554107Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2554297Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2554432Z return mod(**inputs) 2025-09-07T08:07:07.2554679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2554744Z outputs = self.bert( 2025-09-07T08:07:07.2554989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2555059Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2555296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2555358Z layer_outputs = layer_module( 2025-09-07T08:07:07.2555576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2555644Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2555887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2556040Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2556280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2556351Z return func(*args, **kwargs) 2025-09-07T08:07:07.2556591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2556658Z self_outputs = self.self( 2025-09-07T08:07:07.2556893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2556960Z return func(*args, **kwargs) 2025-09-07T08:07:07.2557202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2557321Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2557327Z 2025-09-07T08:07:07.2557431Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2557619Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2557682Z return mod(**inputs) 2025-09-07T08:07:07.2557928Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2557984Z outputs = self.bert( 2025-09-07T08:07:07.2558238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2558300Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2558548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2558610Z layer_outputs = layer_module( 2025-09-07T08:07:07.2558821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2558899Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2559141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2559217Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2559451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2559515Z return func(*args, **kwargs) 2025-09-07T08:07:07.2559754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2559873Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2560119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2560237Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2560272Z 2025-09-07T08:07:07.2560371Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2560556Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2560614Z return mod(**inputs) 2025-09-07T08:07:07.2560865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2560922Z outputs = self.bert( 2025-09-07T08:07:07.2561168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2561232Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2561474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2561542Z layer_outputs = layer_module( 2025-09-07T08:07:07.2561751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2561894Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2562143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2562220Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2562478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2562550Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2562835Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2562943Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2563189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2563263Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2563268Z 2025-09-07T08:07:07.2563365Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2563552Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2563608Z return mod(**inputs) 2025-09-07T08:07:07.2563854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2563911Z outputs = self.bert( 2025-09-07T08:07:07.2564160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2564225Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2564466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2564531Z layer_outputs = layer_module( 2025-09-07T08:07:07.2564746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2564822Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2565061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2565132Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2565394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2565460Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2565738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2565843Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2566093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2566260Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2566469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2566537Z return self.act(input) 2025-09-07T08:07:07.2566540Z 2025-09-07T08:07:07.2566633Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2566824Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2566882Z return mod(**inputs) 2025-09-07T08:07:07.2567138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2567208Z outputs = self.bert( 2025-09-07T08:07:07.2567455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2567528Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2567834Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2567905Z layer_outputs = layer_module( 2025-09-07T08:07:07.2568133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2568208Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2568457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2568536Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2568801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2568874Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2569152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2569291Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2569537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2569616Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2569620Z 2025-09-07T08:07:07.2569718Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2569911Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2569977Z return mod(**inputs) 2025-09-07T08:07:07.2570227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2570290Z outputs = self.bert( 2025-09-07T08:07:07.2570534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2570608Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2570854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2570920Z layer_outputs = layer_module( 2025-09-07T08:07:07.2571141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2571214Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2571462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2571532Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2571776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2571864Z return func(*args, **kwargs) 2025-09-07T08:07:07.2572107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2572245Z self_outputs = self.self( 2025-09-07T08:07:07.2572484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2572547Z return func(*args, **kwargs) 2025-09-07T08:07:07.2572798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2573000Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2573003Z 2025-09-07T08:07:07.2573104Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2573293Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2573358Z return mod(**inputs) 2025-09-07T08:07:07.2573605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2573667Z outputs = self.bert( 2025-09-07T08:07:07.2573975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2574040Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2574289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2574352Z layer_outputs = layer_module( 2025-09-07T08:07:07.2574572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2574647Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2574884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2574961Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2575201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2575266Z return func(*args, **kwargs) 2025-09-07T08:07:07.2575513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2575575Z self_outputs = self.self( 2025-09-07T08:07:07.2575815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2575877Z return func(*args, **kwargs) 2025-09-07T08:07:07.2576122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2576185Z self.key(current_states) 2025-09-07T08:07:07.2576189Z 2025-09-07T08:07:07.2576279Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2576473Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2576535Z return mod(**inputs) 2025-09-07T08:07:07.2576792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2576850Z outputs = self.bert( 2025-09-07T08:07:07.2577092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2577156Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2577394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2577456Z layer_outputs = layer_module( 2025-09-07T08:07:07.2577720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2577785Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2578028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2578191Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2578429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2578487Z return func(*args, **kwargs) 2025-09-07T08:07:07.2578728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2578788Z self_outputs = self.self( 2025-09-07T08:07:07.2579020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2579081Z return func(*args, **kwargs) 2025-09-07T08:07:07.2579318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2579384Z self.value(current_states) 2025-09-07T08:07:07.2579389Z 2025-09-07T08:07:07.2579461Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2579611Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2579800Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2579856Z return mod(**inputs) 2025-09-07T08:07:07.2580103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2580158Z outputs = self.bert( 2025-09-07T08:07:07.2580400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2580463Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2580702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2580764Z layer_outputs = layer_module( 2025-09-07T08:07:07.2580974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2581048Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2581285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2581354Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2581588Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2581645Z return func(*args, **kwargs) 2025-09-07T08:07:07.2581886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2581945Z self_outputs = self.self( 2025-09-07T08:07:07.2582177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2582237Z return func(*args, **kwargs) 2025-09-07T08:07:07.2582477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2582603Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2582606Z 2025-09-07T08:07:07.2582693Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2582876Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2582932Z return mod(**inputs) 2025-09-07T08:07:07.2583174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2583231Z outputs = self.bert( 2025-09-07T08:07:07.2583474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2583537Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2583775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2583904Z layer_outputs = layer_module( 2025-09-07T08:07:07.2584120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2584184Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2584425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2584494Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2584729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2584790Z return func(*args, **kwargs) 2025-09-07T08:07:07.2585028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2585149Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2585447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2585521Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2585527Z 2025-09-07T08:07:07.2585614Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2585797Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2585855Z return mod(**inputs) 2025-09-07T08:07:07.2586098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2586155Z outputs = self.bert( 2025-09-07T08:07:07.2586395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2586457Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2586701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2586763Z layer_outputs = layer_module( 2025-09-07T08:07:07.2586976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2587042Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2587281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2587356Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2587615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2587683Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2587960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2588077Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2588319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2588389Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2588393Z 2025-09-07T08:07:07.2588482Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2588665Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2588721Z return mod(**inputs) 2025-09-07T08:07:07.2588961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2589017Z outputs = self.bert( 2025-09-07T08:07:07.2589259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2589320Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2589618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2589711Z layer_outputs = layer_module( 2025-09-07T08:07:07.2589923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2589991Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2590229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2590302Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2590556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2590623Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2590895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2591083Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2591333Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2591439Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2591648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2591711Z return self.act(input) 2025-09-07T08:07:07.2591714Z 2025-09-07T08:07:07.2591803Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2591987Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2592042Z return mod(**inputs) 2025-09-07T08:07:07.2592287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2592345Z outputs = self.bert( 2025-09-07T08:07:07.2592591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2592654Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2592892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2592955Z layer_outputs = layer_module( 2025-09-07T08:07:07.2593172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2593244Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2593482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2593552Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2593809Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2593877Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2594151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2594273Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2594516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2594588Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2594590Z 2025-09-07T08:07:07.2594678Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2594862Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2594917Z return mod(**inputs) 2025-09-07T08:07:07.2595161Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2595322Z outputs = self.bert( 2025-09-07T08:07:07.2595570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2595634Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2595874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2595941Z layer_outputs = layer_module( 2025-09-07T08:07:07.2596151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2596215Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2596455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2596527Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2596767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2596889Z return func(*args, **kwargs) 2025-09-07T08:07:07.2597132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2597196Z self_outputs = self.self( 2025-09-07T08:07:07.2597431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2597492Z return func(*args, **kwargs) 2025-09-07T08:07:07.2597730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2597928Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2597932Z 2025-09-07T08:07:07.2598019Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2598202Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2598263Z return mod(**inputs) 2025-09-07T08:07:07.2598506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2598564Z outputs = self.bert( 2025-09-07T08:07:07.2598806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2598867Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2599107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2599168Z layer_outputs = layer_module( 2025-09-07T08:07:07.2599380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2599444Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2599686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2599756Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2599989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2600050Z return func(*args, **kwargs) 2025-09-07T08:07:07.2600288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2600349Z self_outputs = self.self( 2025-09-07T08:07:07.2600582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2600643Z return func(*args, **kwargs) 2025-09-07T08:07:07.2600889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2601011Z self.key(current_states) 2025-09-07T08:07:07.2601014Z 2025-09-07T08:07:07.2601143Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2601329Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2601391Z return mod(**inputs) 2025-09-07T08:07:07.2601643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2601700Z outputs = self.bert( 2025-09-07T08:07:07.2601952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2602014Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2602265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2602327Z layer_outputs = layer_module( 2025-09-07T08:07:07.2602539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2602686Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2602931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2603008Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2603429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2603493Z return func(*args, **kwargs) 2025-09-07T08:07:07.2603751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2603814Z self_outputs = self.self( 2025-09-07T08:07:07.2604062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2604122Z return func(*args, **kwargs) 2025-09-07T08:07:07.2604375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2604447Z self.value(current_states) 2025-09-07T08:07:07.2604451Z 2025-09-07T08:07:07.2604523Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2604624Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2604820Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2604884Z return mod(**inputs) 2025-09-07T08:07:07.2605135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2605194Z outputs = self.bert( 2025-09-07T08:07:07.2605447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2605509Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2605759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2605825Z layer_outputs = layer_module( 2025-09-07T08:07:07.2606044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2606121Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2606361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2606437Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2606675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2606743Z return func(*args, **kwargs) 2025-09-07T08:07:07.2606990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2607145Z self_outputs = self.self( 2025-09-07T08:07:07.2607451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2607513Z return func(*args, **kwargs) 2025-09-07T08:07:07.2607759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2607884Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2607887Z 2025-09-07T08:07:07.2607979Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2608169Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2608224Z return mod(**inputs) 2025-09-07T08:07:07.2608474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2608529Z outputs = self.bert( 2025-09-07T08:07:07.2608866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2608935Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2609178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2609242Z layer_outputs = layer_module( 2025-09-07T08:07:07.2609457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2609525Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2609767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2609836Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2610074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2610136Z return func(*args, **kwargs) 2025-09-07T08:07:07.2610381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2610499Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2610737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2610811Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2610814Z 2025-09-07T08:07:07.2610903Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2611092Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2611146Z return mod(**inputs) 2025-09-07T08:07:07.2611389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2611447Z outputs = self.bert( 2025-09-07T08:07:07.2611699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2611768Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2612009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2612074Z layer_outputs = layer_module( 2025-09-07T08:07:07.2612291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2612358Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2612601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2612674Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2612941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2613047Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2613364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2613478Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2613719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2613792Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2613796Z 2025-09-07T08:07:07.2613886Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2614072Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2614130Z return mod(**inputs) 2025-09-07T08:07:07.2614373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2614433Z outputs = self.bert( 2025-09-07T08:07:07.2614736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2614803Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2615045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2615105Z layer_outputs = layer_module( 2025-09-07T08:07:07.2615321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2615387Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2615629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2615698Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2615953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2616025Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2616298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2616404Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2616642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2616742Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2616949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2617009Z return self.act(input) 2025-09-07T08:07:07.2617012Z 2025-09-07T08:07:07.2617104Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2617291Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2617350Z return mod(**inputs) 2025-09-07T08:07:07.2617642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2617699Z outputs = self.bert( 2025-09-07T08:07:07.2617945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2618010Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2618253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2618312Z layer_outputs = layer_module( 2025-09-07T08:07:07.2618527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2618596Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2618837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2618991Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2619250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2619321Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2619595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2619715Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2619962Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2620033Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2620036Z 2025-09-07T08:07:07.2620134Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2620317Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2620457Z return mod(**inputs) 2025-09-07T08:07:07.2620716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2620775Z outputs = self.bert( 2025-09-07T08:07:07.2621022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2621086Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2621329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2621399Z layer_outputs = layer_module( 2025-09-07T08:07:07.2621611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2621686Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2621930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2622013Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2622252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2622316Z return func(*args, **kwargs) 2025-09-07T08:07:07.2622560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2622624Z self_outputs = self.self( 2025-09-07T08:07:07.2622863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2622922Z return func(*args, **kwargs) 2025-09-07T08:07:07.2623163Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2623363Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2623373Z 2025-09-07T08:07:07.2623462Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2623649Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2623705Z return mod(**inputs) 2025-09-07T08:07:07.2623953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2624009Z outputs = self.bert( 2025-09-07T08:07:07.2624252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2624320Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2624559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2624623Z layer_outputs = layer_module( 2025-09-07T08:07:07.2624880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2624981Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2625226Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2625296Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2625533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2625591Z return func(*args, **kwargs) 2025-09-07T08:07:07.2625830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2625893Z self_outputs = self.self( 2025-09-07T08:07:07.2626127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2626192Z return func(*args, **kwargs) 2025-09-07T08:07:07.2626490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2626557Z self.key(current_states) 2025-09-07T08:07:07.2626560Z 2025-09-07T08:07:07.2626648Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2626833Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2626894Z return mod(**inputs) 2025-09-07T08:07:07.2627137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2627195Z outputs = self.bert( 2025-09-07T08:07:07.2627437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2627500Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2627744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2627811Z layer_outputs = layer_module( 2025-09-07T08:07:07.2628027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2628091Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2628333Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2628407Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2628644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2628705Z return func(*args, **kwargs) 2025-09-07T08:07:07.2628943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2629008Z self_outputs = self.self( 2025-09-07T08:07:07.2629244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2629304Z return func(*args, **kwargs) 2025-09-07T08:07:07.2629545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2629606Z self.value(current_states) 2025-09-07T08:07:07.2629609Z 2025-09-07T08:07:07.2629683Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2629768Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2629948Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2630004Z return mod(**inputs) 2025-09-07T08:07:07.2630249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2630310Z outputs = self.bert( 2025-09-07T08:07:07.2630596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2630694Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2630936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2630996Z layer_outputs = layer_module( 2025-09-07T08:07:07.2631213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2631277Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2631517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2631587Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2631820Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2631883Z return func(*args, **kwargs) 2025-09-07T08:07:07.2632183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2632248Z self_outputs = self.self( 2025-09-07T08:07:07.2632479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2632538Z return func(*args, **kwargs) 2025-09-07T08:07:07.2632779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2632900Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2632904Z 2025-09-07T08:07:07.2632994Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2633174Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2633230Z return mod(**inputs) 2025-09-07T08:07:07.2633480Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2633537Z outputs = self.bert( 2025-09-07T08:07:07.2633784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2633843Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2634084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2634146Z layer_outputs = layer_module( 2025-09-07T08:07:07.2634356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2634423Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2634661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2634735Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2634972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2635032Z return func(*args, **kwargs) 2025-09-07T08:07:07.2635271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2635386Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2635630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2635703Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2635705Z 2025-09-07T08:07:07.2635793Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2635979Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2636069Z return mod(**inputs) 2025-09-07T08:07:07.2636319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2636409Z outputs = self.bert( 2025-09-07T08:07:07.2636657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2636719Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2636961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2637024Z layer_outputs = layer_module( 2025-09-07T08:07:07.2637236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2637303Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2637541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2637614Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2637938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2638004Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2638284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2638393Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2638634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2638708Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2638711Z 2025-09-07T08:07:07.2638798Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2638984Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2639040Z return mod(**inputs) 2025-09-07T08:07:07.2639289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2639343Z outputs = self.bert( 2025-09-07T08:07:07.2639585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2639648Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2639889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2639950Z layer_outputs = layer_module( 2025-09-07T08:07:07.2640159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2640224Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2640469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2640545Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2640801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2640864Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2641136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2641238Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2641477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2641577Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2641781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2641882Z return self.act(input) 2025-09-07T08:07:07.2641885Z 2025-09-07T08:07:07.2642006Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2642190Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2642250Z return mod(**inputs) 2025-09-07T08:07:07.2642496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2642554Z outputs = self.bert( 2025-09-07T08:07:07.2642794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2642855Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2643101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2643160Z layer_outputs = layer_module( 2025-09-07T08:07:07.2643373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2643516Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2643765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2643834Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2644086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2644152Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2644424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2644549Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2644792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2644862Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2644870Z 2025-09-07T08:07:07.2644962Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2645142Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2645199Z return mod(**inputs) 2025-09-07T08:07:07.2645448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2645506Z outputs = self.bert( 2025-09-07T08:07:07.2645750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2645811Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2646054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2646115Z layer_outputs = layer_module( 2025-09-07T08:07:07.2646331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2646398Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2646638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2646712Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2646949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2647010Z return func(*args, **kwargs) 2025-09-07T08:07:07.2647250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2647309Z self_outputs = self.self( 2025-09-07T08:07:07.2647546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2647641Z return func(*args, **kwargs) 2025-09-07T08:07:07.2647932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2648128Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2648132Z 2025-09-07T08:07:07.2648225Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2648408Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2648467Z return mod(**inputs) 2025-09-07T08:07:07.2648713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2648769Z outputs = self.bert( 2025-09-07T08:07:07.2649014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2649075Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2649525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2649590Z layer_outputs = layer_module( 2025-09-07T08:07:07.2649802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2649870Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2650110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2650183Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2650417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2650474Z return func(*args, **kwargs) 2025-09-07T08:07:07.2650719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2650793Z self_outputs = self.self( 2025-09-07T08:07:07.2651035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2651092Z return func(*args, **kwargs) 2025-09-07T08:07:07.2651330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2651393Z self.key(current_states) 2025-09-07T08:07:07.2651397Z 2025-09-07T08:07:07.2651490Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2651676Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2651730Z return mod(**inputs) 2025-09-07T08:07:07.2651973Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2652031Z outputs = self.bert( 2025-09-07T08:07:07.2652278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2652344Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2652583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2652644Z layer_outputs = layer_module( 2025-09-07T08:07:07.2652859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2652923Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2653162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2653230Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2653469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2653570Z return func(*args, **kwargs) 2025-09-07T08:07:07.2653846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2653909Z self_outputs = self.self( 2025-09-07T08:07:07.2654142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2654201Z return func(*args, **kwargs) 2025-09-07T08:07:07.2654436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2654498Z self.value(current_states) 2025-09-07T08:07:07.2654501Z 2025-09-07T08:07:07.2654574Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2654660Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2654844Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2654899Z return mod(**inputs) 2025-09-07T08:07:07.2655202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2655267Z outputs = self.bert( 2025-09-07T08:07:07.2655512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2655577Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2655818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2655880Z layer_outputs = layer_module( 2025-09-07T08:07:07.2656089Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2656152Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2656395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2656468Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2656710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2656766Z return func(*args, **kwargs) 2025-09-07T08:07:07.2657007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2657069Z self_outputs = self.self( 2025-09-07T08:07:07.2657303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2657361Z return func(*args, **kwargs) 2025-09-07T08:07:07.2657651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2657770Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2657776Z 2025-09-07T08:07:07.2657864Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2658052Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2658113Z return mod(**inputs) 2025-09-07T08:07:07.2658357Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2658414Z outputs = self.bert( 2025-09-07T08:07:07.2658660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2658721Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2658965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2659023Z layer_outputs = layer_module( 2025-09-07T08:07:07.2659236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2659340Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2659620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2659694Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2659930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2659992Z return func(*args, **kwargs) 2025-09-07T08:07:07.2660228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2660343Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2660589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2660661Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2660667Z 2025-09-07T08:07:07.2660759Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2661004Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2661069Z return mod(**inputs) 2025-09-07T08:07:07.2661312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2661368Z outputs = self.bert( 2025-09-07T08:07:07.2661614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2661675Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2661917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2661979Z layer_outputs = layer_module( 2025-09-07T08:07:07.2662197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2662270Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2662512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2662586Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2662843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2662911Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2663188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2663294Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2663534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2663601Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2663606Z 2025-09-07T08:07:07.2663699Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2663880Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2663935Z return mod(**inputs) 2025-09-07T08:07:07.2664178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2664232Z outputs = self.bert( 2025-09-07T08:07:07.2664475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2664535Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2664772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2664834Z layer_outputs = layer_module( 2025-09-07T08:07:07.2665042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2665197Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2665437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2665507Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2665761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2665825Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2666103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2666205Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2666445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2666547Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2666815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2666881Z return self.act(input) 2025-09-07T08:07:07.2666884Z 2025-09-07T08:07:07.2666973Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2667167Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2667223Z return mod(**inputs) 2025-09-07T08:07:07.2667475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2667532Z outputs = self.bert( 2025-09-07T08:07:07.2667775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2667842Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2668083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2668156Z layer_outputs = layer_module( 2025-09-07T08:07:07.2668371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2668438Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2668687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2668759Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2669022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2669088Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2669362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2669493Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2669752Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2669829Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2669833Z 2025-09-07T08:07:07.2669923Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2670115Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2670173Z return mod(**inputs) 2025-09-07T08:07:07.2670423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2670487Z outputs = self.bert( 2025-09-07T08:07:07.2670734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2670807Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2671087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2671180Z layer_outputs = layer_module( 2025-09-07T08:07:07.2671399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2671468Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2671720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2671791Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2672036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2672100Z return func(*args, **kwargs) 2025-09-07T08:07:07.2672341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2672412Z self_outputs = self.self( 2025-09-07T08:07:07.2672711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2672777Z return func(*args, **kwargs) 2025-09-07T08:07:07.2673020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2673224Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2673227Z 2025-09-07T08:07:07.2673321Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2673505Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2673569Z return mod(**inputs) 2025-09-07T08:07:07.2673906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2673994Z outputs = self.bert( 2025-09-07T08:07:07.2674301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2680883Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2681210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2681280Z layer_outputs = layer_module( 2025-09-07T08:07:07.2681530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2681606Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2681866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2681942Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2682190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2682265Z return func(*args, **kwargs) 2025-09-07T08:07:07.2682518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2682588Z self_outputs = self.self( 2025-09-07T08:07:07.2682824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2682885Z return func(*args, **kwargs) 2025-09-07T08:07:07.2683131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2683193Z self.key(current_states) 2025-09-07T08:07:07.2683197Z 2025-09-07T08:07:07.2683300Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2683500Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2683634Z return mod(**inputs) 2025-09-07T08:07:07.2683894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2683991Z outputs = self.bert( 2025-09-07T08:07:07.2684251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2684320Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2684581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2684647Z layer_outputs = layer_module( 2025-09-07T08:07:07.2684865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2684944Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2685186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2685265Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2685577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2685638Z return func(*args, **kwargs) 2025-09-07T08:07:07.2685881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2685943Z self_outputs = self.self( 2025-09-07T08:07:07.2686182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2686240Z return func(*args, **kwargs) 2025-09-07T08:07:07.2686485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2686549Z self.value(current_states) 2025-09-07T08:07:07.2686552Z 2025-09-07T08:07:07.2686624Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2686726Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2686925Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2686984Z return mod(**inputs) 2025-09-07T08:07:07.2687233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2687291Z outputs = self.bert( 2025-09-07T08:07:07.2687547Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2687609Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2687856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2687919Z layer_outputs = layer_module( 2025-09-07T08:07:07.2688132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2688202Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2688442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2688515Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2688750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2688811Z return func(*args, **kwargs) 2025-09-07T08:07:07.2689054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2689112Z self_outputs = self.self( 2025-09-07T08:07:07.2689350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2689408Z return func(*args, **kwargs) 2025-09-07T08:07:07.2689653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2689847Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2689851Z 2025-09-07T08:07:07.2689943Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2690143Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2690199Z return mod(**inputs) 2025-09-07T08:07:07.2690450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2690506Z outputs = self.bert( 2025-09-07T08:07:07.2690750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2690817Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2691060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2691133Z layer_outputs = layer_module( 2025-09-07T08:07:07.2691407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2691479Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2691724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2691793Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2692031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2692090Z return func(*args, **kwargs) 2025-09-07T08:07:07.2692334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2692454Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2692698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2692776Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2692780Z 2025-09-07T08:07:07.2692869Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2693056Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2693110Z return mod(**inputs) 2025-09-07T08:07:07.2693363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2693421Z outputs = self.bert( 2025-09-07T08:07:07.2693678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2693743Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2693984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2694052Z layer_outputs = layer_module( 2025-09-07T08:07:07.2694270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2694336Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2694576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2694647Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2694911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2694979Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2695256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2695370Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2695653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2695758Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2695761Z 2025-09-07T08:07:07.2695854Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2696047Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2696102Z return mod(**inputs) 2025-09-07T08:07:07.2696349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2696408Z outputs = self.bert( 2025-09-07T08:07:07.2696648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2696711Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2696953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2697076Z layer_outputs = layer_module( 2025-09-07T08:07:07.2697296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2697365Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2697688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2697762Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2698018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2698082Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2698353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2698470Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2698714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2698820Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2699030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2699089Z return self.act(input) 2025-09-07T08:07:07.2699092Z 2025-09-07T08:07:07.2699185Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2699371Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2699429Z return mod(**inputs) 2025-09-07T08:07:07.2699673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2699729Z outputs = self.bert( 2025-09-07T08:07:07.2699986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2700049Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2700292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2700353Z layer_outputs = layer_module( 2025-09-07T08:07:07.2700567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2700634Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2700873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2700944Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2701195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2701305Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2701627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2701754Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2702003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2702072Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2702075Z 2025-09-07T08:07:07.2702172Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2702355Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2702411Z return mod(**inputs) 2025-09-07T08:07:07.2702658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2702716Z outputs = self.bert( 2025-09-07T08:07:07.2703021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2703250Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2703495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2703555Z layer_outputs = layer_module( 2025-09-07T08:07:07.2703766Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2703832Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2704071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2704143Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2704383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2704445Z return func(*args, **kwargs) 2025-09-07T08:07:07.2704695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2704754Z self_outputs = self.self( 2025-09-07T08:07:07.2704990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2705048Z return func(*args, **kwargs) 2025-09-07T08:07:07.2705288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2705495Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2705499Z 2025-09-07T08:07:07.2705587Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2705771Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2705826Z return mod(**inputs) 2025-09-07T08:07:07.2706076Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2706131Z outputs = self.bert( 2025-09-07T08:07:07.2706377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2706446Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2706686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2706754Z layer_outputs = layer_module( 2025-09-07T08:07:07.2706966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2707031Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2707278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2707467Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2707713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2707773Z return func(*args, **kwargs) 2025-09-07T08:07:07.2708020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2708080Z self_outputs = self.self( 2025-09-07T08:07:07.2708315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2708381Z return func(*args, **kwargs) 2025-09-07T08:07:07.2708622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2708689Z self.key(current_states) 2025-09-07T08:07:07.2708693Z 2025-09-07T08:07:07.2708784Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2709068Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2709134Z return mod(**inputs) 2025-09-07T08:07:07.2709393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2709458Z outputs = self.bert( 2025-09-07T08:07:07.2709704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2709767Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2710013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2710075Z layer_outputs = layer_module( 2025-09-07T08:07:07.2710295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2710363Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2710606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2710675Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2710907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2710965Z return func(*args, **kwargs) 2025-09-07T08:07:07.2711201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2711261Z self_outputs = self.self( 2025-09-07T08:07:07.2711492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2711550Z return func(*args, **kwargs) 2025-09-07T08:07:07.2711788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2711853Z self.value(current_states) 2025-09-07T08:07:07.2711857Z 2025-09-07T08:07:07.2711929Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2712019Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2712201Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2712258Z return mod(**inputs) 2025-09-07T08:07:07.2712500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2712556Z outputs = self.bert( 2025-09-07T08:07:07.2712797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2712860Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2713100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2713226Z layer_outputs = layer_module( 2025-09-07T08:07:07.2713445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2713510Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2713751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2713818Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2714053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2714115Z return func(*args, **kwargs) 2025-09-07T08:07:07.2714356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2714415Z self_outputs = self.self( 2025-09-07T08:07:07.2714650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2714772Z return func(*args, **kwargs) 2025-09-07T08:07:07.2715019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2715142Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2715145Z 2025-09-07T08:07:07.2715240Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2715422Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2715477Z return mod(**inputs) 2025-09-07T08:07:07.2715722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2715779Z outputs = self.bert( 2025-09-07T08:07:07.2716025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2716091Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2716329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2716390Z layer_outputs = layer_module( 2025-09-07T08:07:07.2716602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2716671Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2716910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2716982Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2717212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2717270Z return func(*args, **kwargs) 2025-09-07T08:07:07.2717517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2717634Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2717885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2717956Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2717959Z 2025-09-07T08:07:07.2718050Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2718240Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2718297Z return mod(**inputs) 2025-09-07T08:07:07.2718555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2718611Z outputs = self.bert( 2025-09-07T08:07:07.2718859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2718988Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2719236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2719303Z layer_outputs = layer_module( 2025-09-07T08:07:07.2719519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2719599Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2719843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2719919Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2720188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2720255Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2720603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2720723Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2720968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2721043Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2721047Z 2025-09-07T08:07:07.2721140Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2721333Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2721390Z return mod(**inputs) 2025-09-07T08:07:07.2721641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2721698Z outputs = self.bert( 2025-09-07T08:07:07.2721948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2722020Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2722266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2722338Z layer_outputs = layer_module( 2025-09-07T08:07:07.2722553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2722625Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2722874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2722945Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2723210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2723279Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2723559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2723665Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2723903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2724009Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2724219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2724289Z return self.act(input) 2025-09-07T08:07:07.2724292Z 2025-09-07T08:07:07.2724381Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2724567Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2724628Z return mod(**inputs) 2025-09-07T08:07:07.2724931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2725038Z outputs = self.bert( 2025-09-07T08:07:07.2725291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2725355Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2725603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2725666Z layer_outputs = layer_module( 2025-09-07T08:07:07.2725886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2725953Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2726193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2726275Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2726593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2726666Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2726936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2727061Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2727312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2727385Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2727388Z 2025-09-07T08:07:07.2727485Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2727672Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2727737Z return mod(**inputs) 2025-09-07T08:07:07.2727989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2728044Z outputs = self.bert( 2025-09-07T08:07:07.2728290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2728353Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2728599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2728660Z layer_outputs = layer_module( 2025-09-07T08:07:07.2728876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2728950Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2729189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2729266Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2729508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2729570Z return func(*args, **kwargs) 2025-09-07T08:07:07.2729815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2729876Z self_outputs = self.self( 2025-09-07T08:07:07.2730113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2730172Z return func(*args, **kwargs) 2025-09-07T08:07:07.2730420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2730625Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2730664Z 2025-09-07T08:07:07.2730785Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2730973Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2731029Z return mod(**inputs) 2025-09-07T08:07:07.2731276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2731330Z outputs = self.bert( 2025-09-07T08:07:07.2731573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2731641Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2731881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2731948Z layer_outputs = layer_module( 2025-09-07T08:07:07.2732158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2732490Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2732739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2732813Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2733055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2733114Z return func(*args, **kwargs) 2025-09-07T08:07:07.2733357Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2733416Z self_outputs = self.self( 2025-09-07T08:07:07.2733652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2733716Z return func(*args, **kwargs) 2025-09-07T08:07:07.2733959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2734023Z self.key(current_states) 2025-09-07T08:07:07.2734027Z 2025-09-07T08:07:07.2734115Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2734301Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2734364Z return mod(**inputs) 2025-09-07T08:07:07.2734616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2734674Z outputs = self.bert( 2025-09-07T08:07:07.2734916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2734983Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2735231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2735293Z layer_outputs = layer_module( 2025-09-07T08:07:07.2735515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2735579Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2735819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2735890Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2736122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2736186Z return func(*args, **kwargs) 2025-09-07T08:07:07.2736423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2736485Z self_outputs = self.self( 2025-09-07T08:07:07.2736759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2736853Z return func(*args, **kwargs) 2025-09-07T08:07:07.2737099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2737160Z self.value(current_states) 2025-09-07T08:07:07.2737163Z 2025-09-07T08:07:07.2737239Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2737325Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2737506Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2737612Z return mod(**inputs) 2025-09-07T08:07:07.2737858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2737915Z outputs = self.bert( 2025-09-07T08:07:07.2738156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2738286Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2738533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2738598Z layer_outputs = layer_module( 2025-09-07T08:07:07.2738815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2738883Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2739126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2739194Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2739427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2739490Z return func(*args, **kwargs) 2025-09-07T08:07:07.2739730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2739795Z self_outputs = self.self( 2025-09-07T08:07:07.2740028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2740087Z return func(*args, **kwargs) 2025-09-07T08:07:07.2740327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2740448Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2740451Z 2025-09-07T08:07:07.2740546Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2740729Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2740788Z return mod(**inputs) 2025-09-07T08:07:07.2741032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2741092Z outputs = self.bert( 2025-09-07T08:07:07.2741337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2741399Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2741645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2741705Z layer_outputs = layer_module( 2025-09-07T08:07:07.2741917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2741986Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2742226Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2742302Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2742579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2742669Z return func(*args, **kwargs) 2025-09-07T08:07:07.2742914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2743034Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2743277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2743349Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2743353Z 2025-09-07T08:07:07.2743442Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2743626Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2743681Z return mod(**inputs) 2025-09-07T08:07:07.2743928Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2744060Z outputs = self.bert( 2025-09-07T08:07:07.2744314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2744375Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2744615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2744679Z layer_outputs = layer_module( 2025-09-07T08:07:07.2744890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2744957Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2745194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2745269Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2745533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2745602Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2745884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2745995Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2746238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2746308Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2746311Z 2025-09-07T08:07:07.2746400Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2746588Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2746646Z return mod(**inputs) 2025-09-07T08:07:07.2746895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2746953Z outputs = self.bert( 2025-09-07T08:07:07.2747198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2747265Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2747504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2747567Z layer_outputs = layer_module( 2025-09-07T08:07:07.2747778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2747846Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2748086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2748192Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2748495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2748564Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2748847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2748949Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2749188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2749293Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2749500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2749567Z return self.act(input) 2025-09-07T08:07:07.2749570Z 2025-09-07T08:07:07.2749661Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2749910Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2749966Z return mod(**inputs) 2025-09-07T08:07:07.2750215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2750281Z outputs = self.bert( 2025-09-07T08:07:07.2750523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2750590Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2750828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2750889Z layer_outputs = layer_module( 2025-09-07T08:07:07.2751106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2751172Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2751417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2751485Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2751743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2751810Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2752081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2752210Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2752452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2752521Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2752526Z 2025-09-07T08:07:07.2752613Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2752799Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2752860Z return mod(**inputs) 2025-09-07T08:07:07.2753106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2753168Z outputs = self.bert( 2025-09-07T08:07:07.2753412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2753472Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2753713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2753774Z layer_outputs = layer_module( 2025-09-07T08:07:07.2753990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2754145Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2754406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2754481Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2754727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2754793Z return func(*args, **kwargs) 2025-09-07T08:07:07.2755036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2755103Z self_outputs = self.self( 2025-09-07T08:07:07.2755339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2755401Z return func(*args, **kwargs) 2025-09-07T08:07:07.2755713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2755921Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2755925Z 2025-09-07T08:07:07.2756023Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2756212Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2756270Z return mod(**inputs) 2025-09-07T08:07:07.2756522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2756579Z outputs = self.bert( 2025-09-07T08:07:07.2756831Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2756897Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2757143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2757211Z layer_outputs = layer_module( 2025-09-07T08:07:07.2757427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2757501Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2757740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2757816Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2758051Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2758110Z return func(*args, **kwargs) 2025-09-07T08:07:07.2758354Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2758416Z self_outputs = self.self( 2025-09-07T08:07:07.2758656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2758716Z return func(*args, **kwargs) 2025-09-07T08:07:07.2758957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2759023Z self.key(current_states) 2025-09-07T08:07:07.2759026Z 2025-09-07T08:07:07.2759119Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2759308Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2759366Z return mod(**inputs) 2025-09-07T08:07:07.2759616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2759673Z outputs = self.bert( 2025-09-07T08:07:07.2759915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2760049Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2760291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2760358Z layer_outputs = layer_module( 2025-09-07T08:07:07.2760570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2760640Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2760888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2760959Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2761198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2761257Z return func(*args, **kwargs) 2025-09-07T08:07:07.2761566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2761636Z self_outputs = self.self( 2025-09-07T08:07:07.2761882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2761947Z return func(*args, **kwargs) 2025-09-07T08:07:07.2762185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2762252Z self.value(current_states) 2025-09-07T08:07:07.2762255Z 2025-09-07T08:07:07.2762329Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2762420Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2762612Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2762668Z return mod(**inputs) 2025-09-07T08:07:07.2762920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2762978Z outputs = self.bert( 2025-09-07T08:07:07.2763221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2763287Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2763525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2763591Z layer_outputs = layer_module( 2025-09-07T08:07:07.2763802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2763867Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2764110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2764178Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2764421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2764479Z return func(*args, **kwargs) 2025-09-07T08:07:07.2764726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2764786Z self_outputs = self.self( 2025-09-07T08:07:07.2765022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2765084Z return func(*args, **kwargs) 2025-09-07T08:07:07.2765321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2765447Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2765450Z 2025-09-07T08:07:07.2765540Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2765764Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2765855Z return mod(**inputs) 2025-09-07T08:07:07.2766100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2766158Z outputs = self.bert( 2025-09-07T08:07:07.2766400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2766463Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2766702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2766763Z layer_outputs = layer_module( 2025-09-07T08:07:07.2766981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2767046Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2767362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2767434Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2767668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2767729Z return func(*args, **kwargs) 2025-09-07T08:07:07.2767967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2768086Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2768323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2768395Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2768398Z 2025-09-07T08:07:07.2768490Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2768675Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2768735Z return mod(**inputs) 2025-09-07T08:07:07.2768979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2769038Z outputs = self.bert( 2025-09-07T08:07:07.2769284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2769347Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2769591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2769650Z layer_outputs = layer_module( 2025-09-07T08:07:07.2769867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2769934Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2770179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2770257Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2770517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2770589Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2770869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2770982Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2771228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2771299Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2771302Z 2025-09-07T08:07:07.2771436Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2771690Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2771756Z return mod(**inputs) 2025-09-07T08:07:07.2772005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2772062Z outputs = self.bert( 2025-09-07T08:07:07.2772309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2772373Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2772619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2772681Z layer_outputs = layer_module( 2025-09-07T08:07:07.2772895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2772968Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2773270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2773348Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2773605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2773672Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2773954Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2774059Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2774305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2774409Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2774625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2774691Z return self.act(input) 2025-09-07T08:07:07.2774694Z 2025-09-07T08:07:07.2774782Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2774975Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2775033Z return mod(**inputs) 2025-09-07T08:07:07.2775277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2775334Z outputs = self.bert( 2025-09-07T08:07:07.2775578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2775645Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2775884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2775954Z layer_outputs = layer_module( 2025-09-07T08:07:07.2776169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2776239Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2776479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2776549Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2776809Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2776874Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2777147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2777274Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2777605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2777710Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2777713Z 2025-09-07T08:07:07.2777803Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2777991Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2778047Z return mod(**inputs) 2025-09-07T08:07:07.2778296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2778354Z outputs = self.bert( 2025-09-07T08:07:07.2778596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2778663Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2778905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2779033Z layer_outputs = layer_module( 2025-09-07T08:07:07.2779249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2779317Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2779572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2779644Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2779884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2779944Z return func(*args, **kwargs) 2025-09-07T08:07:07.2780182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2780246Z self_outputs = self.self( 2025-09-07T08:07:07.2780484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2780551Z return func(*args, **kwargs) 2025-09-07T08:07:07.2780790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2780998Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2781002Z 2025-09-07T08:07:07.2781093Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2781276Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2781339Z return mod(**inputs) 2025-09-07T08:07:07.2781582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2781645Z outputs = self.bert( 2025-09-07T08:07:07.2781890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2781956Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2782205Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2782267Z layer_outputs = layer_module( 2025-09-07T08:07:07.2782485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2782553Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2782795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2782870Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2783107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2783210Z return func(*args, **kwargs) 2025-09-07T08:07:07.2783491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2783555Z self_outputs = self.self( 2025-09-07T08:07:07.2783794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2783855Z return func(*args, **kwargs) 2025-09-07T08:07:07.2784101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2784162Z self.key(current_states) 2025-09-07T08:07:07.2784165Z 2025-09-07T08:07:07.2784259Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2784442Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2784500Z return mod(**inputs) 2025-09-07T08:07:07.2784750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2784868Z outputs = self.bert( 2025-09-07T08:07:07.2785120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2785185Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2785428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2785494Z layer_outputs = layer_module( 2025-09-07T08:07:07.2785707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2785781Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2786021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2786096Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2786338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2786401Z return func(*args, **kwargs) 2025-09-07T08:07:07.2786648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2786710Z self_outputs = self.self( 2025-09-07T08:07:07.2786947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2787007Z return func(*args, **kwargs) 2025-09-07T08:07:07.2787252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2787321Z self.value(current_states) 2025-09-07T08:07:07.2787325Z 2025-09-07T08:07:07.2787397Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2787494Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2787679Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2787739Z return mod(**inputs) 2025-09-07T08:07:07.2787989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2788048Z outputs = self.bert( 2025-09-07T08:07:07.2788295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2788358Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2788603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2788665Z layer_outputs = layer_module( 2025-09-07T08:07:07.2788874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2788984Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2789277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2789356Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2789593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2789654Z return func(*args, **kwargs) 2025-09-07T08:07:07.2789902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2789963Z self_outputs = self.self( 2025-09-07T08:07:07.2790205Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2790265Z return func(*args, **kwargs) 2025-09-07T08:07:07.2790504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2790690Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2790696Z 2025-09-07T08:07:07.2790785Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2790975Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2791031Z return mod(**inputs) 2025-09-07T08:07:07.2791284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2791342Z outputs = self.bert( 2025-09-07T08:07:07.2791585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2791652Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2791890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2791959Z layer_outputs = layer_module( 2025-09-07T08:07:07.2792176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2792241Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2792487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2792556Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2792796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2792853Z return func(*args, **kwargs) 2025-09-07T08:07:07.2793096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2793219Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2793463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2793552Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2793555Z 2025-09-07T08:07:07.2793641Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2793828Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2793885Z return mod(**inputs) 2025-09-07T08:07:07.2794128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2794191Z outputs = self.bert( 2025-09-07T08:07:07.2794435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2794501Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2794741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2794841Z layer_outputs = layer_module( 2025-09-07T08:07:07.2795095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2795163Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2795412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2795487Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2795746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2795816Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2796095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2796211Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2796450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2796589Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2796593Z 2025-09-07T08:07:07.2796686Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2796885Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2796950Z return mod(**inputs) 2025-09-07T08:07:07.2797208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2797273Z outputs = self.bert( 2025-09-07T08:07:07.2797518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2797585Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2797836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2797901Z layer_outputs = layer_module( 2025-09-07T08:07:07.2798130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2798205Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2798453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2798533Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2798793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2798868Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2799148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2799267Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2799516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2799620Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2799830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2799892Z return self.act(input) 2025-09-07T08:07:07.2799895Z 2025-09-07T08:07:07.2799997Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2800189Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2800250Z return mod(**inputs) 2025-09-07T08:07:07.2800496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2800555Z outputs = self.bert( 2025-09-07T08:07:07.2800803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2800933Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2801177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2801241Z layer_outputs = layer_module( 2025-09-07T08:07:07.2801460Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2801534Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2801775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2801854Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2802110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2802179Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2802519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2802649Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2802897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2802969Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2802972Z 2025-09-07T08:07:07.2803216Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2803407Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2803465Z return mod(**inputs) 2025-09-07T08:07:07.2803722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2803779Z outputs = self.bert( 2025-09-07T08:07:07.2804031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2804099Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2804341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2804405Z layer_outputs = layer_module( 2025-09-07T08:07:07.2804623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2804696Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2804933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2805005Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2805256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2805323Z return func(*args, **kwargs) 2025-09-07T08:07:07.2805568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2805631Z self_outputs = self.self( 2025-09-07T08:07:07.2805875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2805935Z return func(*args, **kwargs) 2025-09-07T08:07:07.2806176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2806379Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2806383Z 2025-09-07T08:07:07.2806473Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2806665Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2806721Z return mod(**inputs) 2025-09-07T08:07:07.2807037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2807315Z outputs = self.bert( 2025-09-07T08:07:07.2807573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2807644Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2807888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2807958Z layer_outputs = layer_module( 2025-09-07T08:07:07.2808173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2808240Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2808486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2808559Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2808905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2808967Z return func(*args, **kwargs) 2025-09-07T08:07:07.2809211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2809281Z self_outputs = self.self( 2025-09-07T08:07:07.2809520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2809585Z return func(*args, **kwargs) 2025-09-07T08:07:07.2809825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2809888Z self.key(current_states) 2025-09-07T08:07:07.2809899Z 2025-09-07T08:07:07.2809993Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2810188Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2810254Z return mod(**inputs) 2025-09-07T08:07:07.2810502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2810565Z outputs = self.bert( 2025-09-07T08:07:07.2810810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2810874Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2811123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2811186Z layer_outputs = layer_module( 2025-09-07T08:07:07.2811406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2811472Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2811716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2811796Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2812032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2812098Z return func(*args, **kwargs) 2025-09-07T08:07:07.2812337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2812398Z self_outputs = self.self( 2025-09-07T08:07:07.2812640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2812701Z return func(*args, **kwargs) 2025-09-07T08:07:07.2812948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2813066Z self.value(current_states) 2025-09-07T08:07:07.2813099Z 2025-09-07T08:07:07.2813180Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2813272Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2813459Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2813524Z return mod(**inputs) 2025-09-07T08:07:07.2813769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2813831Z outputs = self.bert( 2025-09-07T08:07:07.2814077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2814142Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2814389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2814453Z layer_outputs = layer_module( 2025-09-07T08:07:07.2814736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2814805Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2815046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2815124Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2815358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2815422Z return func(*args, **kwargs) 2025-09-07T08:07:07.2815661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2815723Z self_outputs = self.self( 2025-09-07T08:07:07.2815971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2816041Z return func(*args, **kwargs) 2025-09-07T08:07:07.2816297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2816424Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2816427Z 2025-09-07T08:07:07.2816530Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2816721Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2816783Z return mod(**inputs) 2025-09-07T08:07:07.2817041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2817099Z outputs = self.bert( 2025-09-07T08:07:07.2817350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2817421Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2817742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2817818Z layer_outputs = layer_module( 2025-09-07T08:07:07.2818040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2818116Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2818359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2818437Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2818678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2818740Z return func(*args, **kwargs) 2025-09-07T08:07:07.2818988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2819181Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2819430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2819506Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2819509Z 2025-09-07T08:07:07.2819605Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2819804Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2819862Z return mod(**inputs) 2025-09-07T08:07:07.2820115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2820173Z outputs = self.bert( 2025-09-07T08:07:07.2820419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2820492Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2820795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2820863Z layer_outputs = layer_module( 2025-09-07T08:07:07.2821077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2821150Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2821391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2821467Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2821734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2821805Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2822088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2822206Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2822449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2822528Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2822531Z 2025-09-07T08:07:07.2822622Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2822816Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2822872Z return mod(**inputs) 2025-09-07T08:07:07.2823123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2823181Z outputs = self.bert( 2025-09-07T08:07:07.2823425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2823498Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2823740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2823810Z layer_outputs = layer_module( 2025-09-07T08:07:07.2824032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2824103Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2824348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2824424Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2824686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2824752Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2825067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2825208Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2825448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2825555Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2825766Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2825834Z return self.act(input) 2025-09-07T08:07:07.2825837Z 2025-09-07T08:07:07.2825929Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2826116Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2826180Z return mod(**inputs) 2025-09-07T08:07:07.2826426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2826547Z outputs = self.bert( 2025-09-07T08:07:07.2826797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2826863Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2827110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2827173Z layer_outputs = layer_module( 2025-09-07T08:07:07.2827388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2827461Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2827701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2827780Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2828039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2828109Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2828383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2828512Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2828751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2828822Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2828825Z 2025-09-07T08:07:07.2828920Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2829105Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2829163Z return mod(**inputs) 2025-09-07T08:07:07.2829413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2829471Z outputs = self.bert( 2025-09-07T08:07:07.2829718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2829780Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2830022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2830084Z layer_outputs = layer_module( 2025-09-07T08:07:07.2830304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2830369Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2830610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2830720Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2830997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2831061Z return func(*args, **kwargs) 2025-09-07T08:07:07.2831304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2831368Z self_outputs = self.self( 2025-09-07T08:07:07.2831610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2831672Z return func(*args, **kwargs) 2025-09-07T08:07:07.2831916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 374, in forward 2025-09-07T08:07:07.2832116Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:07:07.2832122Z 2025-09-07T08:07:07.2832222Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2832486Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2832547Z return mod(**inputs) 2025-09-07T08:07:07.2832802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2832861Z outputs = self.bert( 2025-09-07T08:07:07.2833110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2833176Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2833418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2833484Z layer_outputs = layer_module( 2025-09-07T08:07:07.2833698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2833781Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2834027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2834099Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2834338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2834400Z return func(*args, **kwargs) 2025-09-07T08:07:07.2834644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2834708Z self_outputs = self.self( 2025-09-07T08:07:07.2834946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2835007Z return func(*args, **kwargs) 2025-09-07T08:07:07.2835249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 402, in forward 2025-09-07T08:07:07.2835323Z self.key(current_states) 2025-09-07T08:07:07.2835326Z 2025-09-07T08:07:07.2835418Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2835611Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2835670Z return mod(**inputs) 2025-09-07T08:07:07.2835916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2835974Z outputs = self.bert( 2025-09-07T08:07:07.2836218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2836289Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2836529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2836630Z layer_outputs = layer_module( 2025-09-07T08:07:07.2836881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2836949Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2837194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2837264Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2837507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2837566Z return func(*args, **kwargs) 2025-09-07T08:07:07.2837805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2837868Z self_outputs = self.self( 2025-09-07T08:07:07.2838099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2838162Z return func(*args, **kwargs) 2025-09-07T08:07:07.2838464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 407, in forward 2025-09-07T08:07:07.2838529Z self.value(current_states) 2025-09-07T08:07:07.2838532Z 2025-09-07T08:07:07.2838608Z cudagraph partition due to non gpu ops 2025-09-07T08:07:07.2838700Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2838891Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2838946Z return mod(**inputs) 2025-09-07T08:07:07.2839191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2839257Z outputs = self.bert( 2025-09-07T08:07:07.2839500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2839569Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2839812Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2839873Z layer_outputs = layer_module( 2025-09-07T08:07:07.2840085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2840150Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2840396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2840465Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2840705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2840764Z return func(*args, **kwargs) 2025-09-07T08:07:07.2841004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 514, in forward 2025-09-07T08:07:07.2841070Z self_outputs = self.self( 2025-09-07T08:07:07.2841306Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2841373Z return func(*args, **kwargs) 2025-09-07T08:07:07.2841611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 438, in forward 2025-09-07T08:07:07.2841736Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:07:07.2841739Z 2025-09-07T08:07:07.2841834Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2842018Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2842078Z return mod(**inputs) 2025-09-07T08:07:07.2842323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2842423Z outputs = self.bert( 2025-09-07T08:07:07.2842704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2842769Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2843014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2843076Z layer_outputs = layer_module( 2025-09-07T08:07:07.2843291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2843361Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2843602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 584, in forward 2025-09-07T08:07:07.2843678Z self_attention_outputs = self.attention( 2025-09-07T08:07:07.2843913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:07:07.2844044Z return func(*args, **kwargs) 2025-09-07T08:07:07.2844291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 524, in forward 2025-09-07T08:07:07.2844410Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:07:07.2844653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 461, in forward 2025-09-07T08:07:07.2844730Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2844733Z 2025-09-07T08:07:07.2844826Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2845009Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2845071Z return mod(**inputs) 2025-09-07T08:07:07.2845316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2845376Z outputs = self.bert( 2025-09-07T08:07:07.2845625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2845687Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2845932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2845994Z layer_outputs = layer_module( 2025-09-07T08:07:07.2846209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2846285Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2846535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2846619Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2846877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2846953Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2847241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2847357Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2847604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 539, in forward 2025-09-07T08:07:07.2847676Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2847679Z 2025-09-07T08:07:07.2847773Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2847956Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2848013Z return mod(**inputs) 2025-09-07T08:07:07.2848268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2848395Z outputs = self.bert( 2025-09-07T08:07:07.2848647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2848714Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2848956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2849026Z layer_outputs = layer_module( 2025-09-07T08:07:07.2849243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2849317Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2849556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2849628Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2849948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2850016Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2850296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 622, in feed_forward_chunk 2025-09-07T08:07:07.2850399Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:07:07.2850648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 540, in forward 2025-09-07T08:07:07.2850753Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:07:07.2850962Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:07:07.2851032Z return self.act(input) 2025-09-07T08:07:07.2851036Z 2025-09-07T08:07:07.2851131Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2851326Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2851385Z return mod(**inputs) 2025-09-07T08:07:07.2851633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1308, in forward 2025-09-07T08:07:07.2851699Z outputs = self.bert( 2025-09-07T08:07:07.2851947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1028, in forward 2025-09-07T08:07:07.2852019Z encoder_outputs = self.encoder( 2025-09-07T08:07:07.2852280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 675, in forward 2025-09-07T08:07:07.2852348Z layer_outputs = layer_module( 2025-09-07T08:07:07.2852581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:07:07.2852663Z return super().__call__(*args, **kwargs) 2025-09-07T08:07:07.2852917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 614, in forward 2025-09-07T08:07:07.2852992Z layer_output = apply_chunking_to_forward( 2025-09-07T08:07:07.2853251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:07:07.2853318Z return forward_fn(*input_tensors) 2025-09-07T08:07:07.2853590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 623, in feed_forward_chunk 2025-09-07T08:07:07.2853720Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:07:07.2853960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 552, in forward 2025-09-07T08:07:07.2854038Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2854041Z 2025-09-07T08:07:07.2854166Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2854393Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2854452Z return mod(**inputs) 2025-09-07T08:07:07.2854700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1323, in forward 2025-09-07T08:07:07.2854797Z prediction_scores = self.cls(sequence_output) 2025-09-07T08:07:07.2855040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 780, in forward 2025-09-07T08:07:07.2855149Z prediction_scores = self.predictions(sequence_output) 2025-09-07T08:07:07.2855387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 769, in forward 2025-09-07T08:07:07.2855465Z hidden_states = self.transform(hidden_states) 2025-09-07T08:07:07.2855707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 745, in forward 2025-09-07T08:07:07.2855857Z hidden_states = self.dense(hidden_states) 2025-09-07T08:07:07.2855860Z 2025-09-07T08:07:07.2855951Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:07.2856136Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:07:07.2856201Z return mod(**inputs) 2025-09-07T08:07:07.2856448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 1323, in forward 2025-09-07T08:07:07.2856528Z prediction_scores = self.cls(sequence_output) 2025-09-07T08:07:07.2856776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 780, in forward 2025-09-07T08:07:07.2856874Z prediction_scores = self.predictions(sequence_output) 2025-09-07T08:07:07.2857124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/bert/modeling_bert.py", line 770, in forward 2025-09-07T08:07:07.2857209Z hidden_states = self.decoder(hidden_states) 2025-09-07T08:07:07.2857216Z 2025-09-07T08:07:18.3833431Z pass 2025-09-07T08:07:18.3833846Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:07:21.2322369Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:07:21.2323268Z import pynvml # type: ignore[import] 2025-09-07T08:07:23.2406647Z 2025-09-07T08:07:24.6385181Z loading model: 0it [00:00, ?it/s]BigBirdForMaskedLM has generative capabilities, as `prepare_inputs_for_generation` is explicitly defined. However, it doesn't directly inherit from `GenerationMixin`. From 👉v4.50👈 onwards, `PreTrainedModel` will NOT inherit from `GenerationMixin`, and this model will lose the ability to call `generate` and other related functions. 2025-09-07T08:07:24.6386710Z - If you're using `trust_remote_code=True`, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes 2025-09-07T08:07:24.6387536Z - If you are the owner of the model architecture code, please modify your model class such that it inherits from `GenerationMixin` (after `PreTrainedModel`, otherwise you'll get an exception). 2025-09-07T08:07:24.6388187Z - If you are not the owner of the model architecture class, please contact the model code owner to update it. 2025-09-07T08:07:24.6389472Z WARNING:transformers.modeling_utils:BigBirdForMaskedLM has generative capabilities, as `prepare_inputs_for_generation` is explicitly defined. However, it doesn't directly inherit from `GenerationMixin`. From 👉v4.50👈 onwards, `PreTrainedModel` will NOT inherit from `GenerationMixin`, and this model will lose the ability to call `generate` and other related functions. 2025-09-07T08:07:24.6391467Z - If you're using `trust_remote_code=True`, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes 2025-09-07T08:07:24.6392276Z - If you are the owner of the model architecture code, please modify your model class such that it inherits from `GenerationMixin` (after `PreTrainedModel`, otherwise you'll get an exception). 2025-09-07T08:07:24.6392955Z - If you are not the owner of the model architecture class, please contact the model code owner to update it. 2025-09-07T08:07:26.5206389Z Input ids are automatically padded from 819 to 832 to be a multiple of `config.block_size`: 64 2025-09-07T08:07:26.5207091Z WARNING:transformers.models.big_bird.modeling_big_bird:Input ids are automatically padded from 819 to 832 to be a multiple of `config.block_size`: 64 2025-09-07T08:07:28.3347692Z 2025-09-07T08:07:28.3348162Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:07:28.3611749Z cpu eval hf_BigBird 2025-09-07T08:07:30.9311234Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:07:33.9687729Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:07:36.9227343Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:07:44.4860237Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:44.4860800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1897, in forward 2025-09-07T08:07:44.4861341Z attention_mask = torch.ones(((batch_size, seq_length + past_key_values_length)), device=device) 2025-09-07T08:07:44.4861590Z 2025-09-07T08:07:45.7682436Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:45.7683122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2090, in torch_dynamo_resume_in__pad_to_block_size_at_2085 2025-09-07T08:07:45.7683759Z input_ids = nn.functional.pad(input_ids, (0, padding_len), value=pad_token_id) 2025-09-07T08:07:45.7684177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:07:45.7684546Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:07:45.7684712Z 2025-09-07T08:07:45.7684818Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:45.7685356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2103, in torch_dynamo_resume_in__pad_to_block_size_at_2085 2025-09-07T08:07:45.7685853Z attention_mask = nn.functional.pad( 2025-09-07T08:07:45.7686178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:07:45.7686528Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:07:45.7686683Z 2025-09-07T08:07:45.7686781Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:45.7687292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2106, in torch_dynamo_resume_in__pad_to_block_size_at_2085 2025-09-07T08:07:45.7687907Z token_type_ids = nn.functional.pad(token_type_ids, (0, padding_len), value=0) # pad with token_type_id = 0 2025-09-07T08:07:45.7688351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:07:45.7697571Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:07:45.7697769Z 2025-09-07T08:07:48.0221758Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:48.0222379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1983, in torch_dynamo_resume_in_forward_at_1935 2025-09-07T08:07:48.0223247Z embedding_output = self.embeddings( 2025-09-07T08:07:48.0223845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 295, in forward 2025-09-07T08:07:48.0224266Z embeddings = self.LayerNorm(embeddings) 2025-09-07T08:07:48.0224414Z 2025-09-07T08:07:48.0224497Z cudagraph partition due to non gpu ops 2025-09-07T08:07:48.0224716Z cudagraph partition due to non gpu ops 2025-09-07T08:07:48.0224914Z cudagraph partition due to non gpu ops 2025-09-07T08:07:48.0225138Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:48.0225680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1947, in torch_dynamo_resume_in_forward_at_1935 2025-09-07T08:07:48.0226273Z blocked_encoder_mask, band_mask, from_mask, to_mask = self.create_masks_for_block_sparse_attn( 2025-09-07T08:07:48.0226854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2060, in create_masks_for_block_sparse_attn 2025-09-07T08:07:48.0227582Z band_mask = create_band_mask_from_inputs(blocked_encoder_mask, blocked_encoder_mask) 2025-09-07T08:07:48.0228133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2052, in create_band_mask_from_inputs 2025-09-07T08:07:48.0228582Z exp_blocked_to_pad = torch.cat( 2025-09-07T08:07:48.0228721Z 2025-09-07T08:07:48.0228828Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:48.0229352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1947, in torch_dynamo_resume_in_forward_at_1935 2025-09-07T08:07:48.0229973Z blocked_encoder_mask, band_mask, from_mask, to_mask = self.create_masks_for_block_sparse_attn( 2025-09-07T08:07:48.0230567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2060, in create_masks_for_block_sparse_attn 2025-09-07T08:07:48.0231130Z band_mask = create_band_mask_from_inputs(blocked_encoder_mask, blocked_encoder_mask) 2025-09-07T08:07:48.0231674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2055, in create_band_mask_from_inputs 2025-09-07T08:07:48.0232219Z band_mask = torch.einsum("blq,blk->blqk", from_blocked_mask[:, 2:-2], exp_blocked_to_pad) 2025-09-07T08:07:48.0232437Z 2025-09-07T08:07:52.5012081Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:52.5012624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1362, in forward 2025-09-07T08:07:52.5013080Z band_mask = band_mask.to(hidden_states.dtype) 2025-09-07T08:07:52.5013232Z 2025-09-07T08:07:52.5013343Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:52.5013790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1364, in forward 2025-09-07T08:07:52.5014276Z from_mask = from_mask.to(hidden_states.dtype) 2025-09-07T08:07:52.5014443Z 2025-09-07T08:07:52.5014540Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:52.5014966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1366, in forward 2025-09-07T08:07:52.5015376Z to_mask = to_mask.to(hidden_states.dtype) 2025-09-07T08:07:52.5015510Z 2025-09-07T08:07:54.2841811Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:54.2842382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:07:54.2842805Z self.query(hidden_states) 2025-09-07T08:07:54.2842924Z 2025-09-07T08:07:54.2843043Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:54.2845429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 445, in forward 2025-09-07T08:07:54.2845946Z self.key(hidden_states) 2025-09-07T08:07:54.2846068Z 2025-09-07T08:07:54.2846170Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:54.2846603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 450, in forward 2025-09-07T08:07:54.2847002Z self.value(hidden_states) 2025-09-07T08:07:54.2847115Z 2025-09-07T08:07:55.2900555Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:55.2901115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1139, in _bigbird_block_rand_mask_with_head 2025-09-07T08:07:55.2901671Z plan_block_length = np.array(plan_from_length) // from_block_size 2025-09-07T08:07:55.2901861Z 2025-09-07T08:07:57.6446900Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6447977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 602, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6448648Z first_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, 0], key_layer, ndim=4) 2025-09-07T08:07:57.6449204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 494, in torch_bmm_nd_transpose 2025-09-07T08:07:57.6449747Z inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:]).transpose(1, 2) 2025-09-07T08:07:57.6449974Z 2025-09-07T08:07:57.6450083Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6450633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 602, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6451248Z first_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, 0], key_layer, ndim=4) 2025-09-07T08:07:57.6451793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 494, in torch_bmm_nd_transpose 2025-09-07T08:07:57.6452309Z inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:]).transpose(1, 2) 2025-09-07T08:07:57.6452529Z 2025-09-07T08:07:57.6452624Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6453195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 611, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6453803Z first_context_layer = self.torch_bmm_nd(first_attn_weights, value_layer, ndim=4) 2025-09-07T08:07:57.6454291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:07:57.6454818Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:07:57.6455066Z 2025-09-07T08:07:57.6455147Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6455365Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6455570Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6455759Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6455955Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6456150Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6456343Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6456528Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6456720Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6456912Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6457105Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6457292Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6457580Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6457859Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6458138Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6458338Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6458564Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6459104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 642, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6459724Z second_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, 1], second_key_mat, ndim=4) 2025-09-07T08:07:57.6460281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 494, in torch_bmm_nd_transpose 2025-09-07T08:07:57.6460800Z inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:]).transpose(1, 2) 2025-09-07T08:07:57.6461015Z 2025-09-07T08:07:57.6461096Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6461291Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6462905Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6463459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 647, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6464028Z to_mask.new_ones([bsz, 1, 1, n_rand_blocks * to_block_size]), 2025-09-07T08:07:57.6464199Z 2025-09-07T08:07:57.6464303Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6464841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 653, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6465411Z rand_mask.new_ones([bsz, n_heads, from_block_size, 4 * to_block_size]), 2025-09-07T08:07:57.6465610Z 2025-09-07T08:07:57.6465687Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6465890Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6466116Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6466643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 698, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6467138Z first_band_product = torch.einsum( 2025-09-07T08:07:57.6467264Z 2025-09-07T08:07:57.6467332Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6467523Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6467726Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6467934Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6468462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 704, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6468967Z last_band_product = torch.einsum( 2025-09-07T08:07:57.6469093Z 2025-09-07T08:07:57.6469172Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6469370Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6469556Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6469752Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6469945Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6470136Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6470322Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6470566Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6471106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 776, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6471790Z second_last_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, -2], second_last_key_mat, ndim=4) 2025-09-07T08:07:57.6472375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 494, in torch_bmm_nd_transpose 2025-09-07T08:07:57.6473026Z inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:]).transpose(1, 2) 2025-09-07T08:07:57.6473253Z 2025-09-07T08:07:57.6473333Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6473538Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6473764Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6474315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 781, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6474869Z to_mask.new_ones([bsz, 1, 1, n_rand_blocks * to_block_size]), 2025-09-07T08:07:57.6475043Z 2025-09-07T08:07:57.6475141Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6475671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 787, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6476327Z rand_mask.new_ones([bsz, n_heads, from_block_size, 4 * to_block_size]), 2025-09-07T08:07:57.6476520Z 2025-09-07T08:07:57.6476602Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6476794Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6477010Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6477541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 807, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6478157Z last_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, -1], key_layer, ndim=4) 2025-09-07T08:07:57.6478703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 494, in torch_bmm_nd_transpose 2025-09-07T08:07:57.6479219Z inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:]).transpose(1, 2) 2025-09-07T08:07:57.6479444Z 2025-09-07T08:07:57.6479538Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6480062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 807, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6480667Z last_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, -1], key_layer, ndim=4) 2025-09-07T08:07:57.6481201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 494, in torch_bmm_nd_transpose 2025-09-07T08:07:57.6481713Z inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:]).transpose(1, 2) 2025-09-07T08:07:57.6481924Z 2025-09-07T08:07:57.6482014Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6482538Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 813, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6483126Z last_context_layer = self.torch_bmm_nd(last_attn_weights, value_layer, ndim=4) 2025-09-07T08:07:57.6483616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:07:57.6484128Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:07:57.6484358Z 2025-09-07T08:07:57.6484450Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6484975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 602, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6485574Z first_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, 0], key_layer, ndim=4) 2025-09-07T08:07:57.6486166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:07:57.6486640Z return torch.bmm( 2025-09-07T08:07:57.6486738Z 2025-09-07T08:07:57.6486832Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6487373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 611, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6487988Z first_context_layer = self.torch_bmm_nd(first_attn_weights, value_layer, ndim=4) 2025-09-07T08:07:57.6488524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:07:57.6489095Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:07:57.6489336Z 2025-09-07T08:07:57.6489438Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6490083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 574, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6490606Z rand_attn = np.stack(rand_attn, axis=0) 2025-09-07T08:07:57.6490749Z 2025-09-07T08:07:57.6490843Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6491380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 577, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6491945Z rand_attn = torch.cat([rand_attn for _ in range(batch_size)], dim=0) 2025-09-07T08:07:57.6492129Z 2025-09-07T08:07:57.6492212Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6492433Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6492959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 620, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6493466Z second_key_mat = torch.cat( 2025-09-07T08:07:57.6493581Z 2025-09-07T08:07:57.6493679Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6494205Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 642, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6494824Z second_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, 1], second_key_mat, ndim=4) 2025-09-07T08:07:57.6495383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:07:57.6495811Z return torch.bmm( 2025-09-07T08:07:57.6495901Z 2025-09-07T08:07:57.6496002Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6496543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 643, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6497044Z second_seq_pad = torch.cat( 2025-09-07T08:07:57.6497159Z 2025-09-07T08:07:57.6497250Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6497855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 579, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6498390Z rand_mask = self._create_rand_mask_from_inputs( 2025-09-07T08:07:57.6498861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in _create_rand_mask_from_inputs 2025-09-07T08:07:57.6499395Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:07:57.6499887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in 2025-09-07T08:07:57.6500465Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:07:57.6500674Z 2025-09-07T08:07:57.6500779Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6501311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 579, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6501825Z rand_mask = self._create_rand_mask_from_inputs( 2025-09-07T08:07:57.6502280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in _create_rand_mask_from_inputs 2025-09-07T08:07:57.6502798Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:07:57.6503449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in 2025-09-07T08:07:57.6504058Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:07:57.6504275Z 2025-09-07T08:07:57.6504385Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6504920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 579, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6505442Z rand_mask = self._create_rand_mask_from_inputs( 2025-09-07T08:07:57.6505903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in _create_rand_mask_from_inputs 2025-09-07T08:07:57.6506427Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:07:57.6506918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in 2025-09-07T08:07:57.6507406Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:07:57.6507622Z 2025-09-07T08:07:57.6507722Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6508264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 579, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6508798Z rand_mask = self._create_rand_mask_from_inputs( 2025-09-07T08:07:57.6509264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in _create_rand_mask_from_inputs 2025-09-07T08:07:57.6509787Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:07:57.6510277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in 2025-09-07T08:07:57.6510763Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:07:57.6510982Z 2025-09-07T08:07:57.6511076Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6511605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 579, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6512123Z rand_mask = self._create_rand_mask_from_inputs( 2025-09-07T08:07:57.6512567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 994, in _create_rand_mask_from_inputs 2025-09-07T08:07:57.6513125Z rand_mask = torch.stack([p1[i1.flatten()] for p1, i1 in zip(to_blocked_mask, rand_attn)]) 2025-09-07T08:07:57.6513331Z 2025-09-07T08:07:57.6513416Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6513635Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6514223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 651, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6514787Z second_rand_pad = torch.cat( 2025-09-07T08:07:57.6514906Z 2025-09-07T08:07:57.6515007Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6515528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 630, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6516029Z second_value_mat = torch.cat( 2025-09-07T08:07:57.6516156Z 2025-09-07T08:07:57.6516247Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6516771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 665, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6517371Z second_context_layer = self.torch_bmm_nd(second_attn_weights, second_value_mat, ndim=4) 2025-09-07T08:07:57.6517949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:07:57.6518468Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:07:57.6518709Z 2025-09-07T08:07:57.6518804Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6519334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 698, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6519848Z first_band_product = torch.einsum( 2025-09-07T08:07:57.6519976Z 2025-09-07T08:07:57.6520077Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6520620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 676, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6521149Z exp_blocked_key_matrix = torch.cat( 2025-09-07T08:07:57.6521286Z 2025-09-07T08:07:57.6521383Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6521915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 687, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6522552Z inner_band_product = self.torch_bmm_nd_transpose(middle_query_matrix, exp_blocked_key_matrix, ndim=5) 2025-09-07T08:07:57.6523111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:07:57.6523538Z return torch.bmm( 2025-09-07T08:07:57.6523635Z 2025-09-07T08:07:57.6523727Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6524258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 693, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6524891Z rand_band_product = self.torch_bmm_nd_transpose(middle_query_matrix, gathered_key[:, :, 1:-1], ndim=5) 2025-09-07T08:07:57.6525440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:07:57.6525858Z return torch.bmm( 2025-09-07T08:07:57.6525954Z 2025-09-07T08:07:57.6526046Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6526570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 704, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6527077Z last_band_product = torch.einsum( 2025-09-07T08:07:57.6527199Z 2025-09-07T08:07:57.6527298Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6527889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 711, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6528487Z first_band_product += (1.0 - to_mask[:, :, :, :to_block_size].unsqueeze(3)) * attn_mask_penalty 2025-09-07T08:07:57.6528709Z 2025-09-07T08:07:57.6528802Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6529328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 710, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6529877Z inner_band_product += (1.0 - band_mask) * attn_mask_penalty 2025-09-07T08:07:57.6530047Z 2025-09-07T08:07:57.6530137Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6530659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 713, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6531290Z rand_band_product += (1.0 - rand_mask[:, :, 1:-1]) * attn_mask_penalty 2025-09-07T08:07:57.6531479Z 2025-09-07T08:07:57.6531572Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6532094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 712, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6532678Z last_band_product += (1.0 - to_mask[:, :, :, -to_block_size:].unsqueeze(3)) * attn_mask_penalty 2025-09-07T08:07:57.6532886Z 2025-09-07T08:07:57.6532975Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6533500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 716, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6533999Z band_product = torch.cat( 2025-09-07T08:07:57.6534111Z 2025-09-07T08:07:57.6534219Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6534748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 679, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6535252Z exp_blocked_value_matrix = torch.cat( 2025-09-07T08:07:57.6535390Z 2025-09-07T08:07:57.6535482Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6536006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 727, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6536508Z context_layer = self.torch_bmm_nd( 2025-09-07T08:07:57.6536918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:07:57.6537439Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:07:57.6537736Z 2025-09-07T08:07:57.6537832Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6538359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 734, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6538861Z context_layer += self.torch_bmm_nd( 2025-09-07T08:07:57.6539263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:07:57.6539766Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:07:57.6540001Z 2025-09-07T08:07:57.6540092Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6540615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 740, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6541246Z context_layer += torch.einsum( 2025-09-07T08:07:57.6541372Z 2025-09-07T08:07:57.6541478Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6542022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 743, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6542516Z context_layer += torch.einsum( 2025-09-07T08:07:57.6542640Z 2025-09-07T08:07:57.6542734Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6543260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 754, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6543763Z second_last_key_mat = torch.cat( 2025-09-07T08:07:57.6543884Z 2025-09-07T08:07:57.6543982Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6544576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 776, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6545254Z second_last_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, -2], second_last_key_mat, ndim=4) 2025-09-07T08:07:57.6545830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:07:57.6546257Z return torch.bmm( 2025-09-07T08:07:57.6546348Z 2025-09-07T08:07:57.6546448Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6546966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 777, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6547466Z second_last_seq_pad = torch.cat( 2025-09-07T08:07:57.6547594Z 2025-09-07T08:07:57.6547668Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6547893Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6548420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 785, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6548916Z second_last_rand_pad = torch.cat( 2025-09-07T08:07:57.6549047Z 2025-09-07T08:07:57.6549143Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6549664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 764, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6550166Z second_last_value_mat = torch.cat( 2025-09-07T08:07:57.6550287Z 2025-09-07T08:07:57.6550385Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6550902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 799, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6551529Z second_last_context_layer = self.torch_bmm_nd(second_last_attn_weights, second_last_value_mat, ndim=4) 2025-09-07T08:07:57.6552061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:07:57.6552567Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:07:57.6552802Z 2025-09-07T08:07:57.6552901Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6553421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 807, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6554019Z last_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, -1], key_layer, ndim=4) 2025-09-07T08:07:57.6554640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T08:07:57.6555069Z return torch.bmm( 2025-09-07T08:07:57.6555164Z 2025-09-07T08:07:57.6555270Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6555797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 813, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6556366Z last_context_layer = self.torch_bmm_nd(last_attn_weights, value_layer, ndim=4) 2025-09-07T08:07:57.6556854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T08:07:57.6557362Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T08:07:57.6557593Z 2025-09-07T08:07:57.6557676Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6557984Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6558198Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6558725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 743, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6559225Z context_layer += torch.einsum( 2025-09-07T08:07:57.6559347Z 2025-09-07T08:07:57.6559424Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6559618Z cudagraph partition due to non gpu ops 2025-09-07T08:07:57.6559827Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6560361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 817, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6560865Z context_layer = torch.cat( 2025-09-07T08:07:57.6560980Z 2025-09-07T08:07:57.6561084Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:07:57.6561618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 821, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T08:07:57.6562196Z context_layer = context_layer.view((bsz, n_heads, from_seq_len, -1)) * from_mask 2025-09-07T08:07:57.6562411Z 2025-09-07T08:08:07.5210818Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:07.5211421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 478, in torch_dynamo_resume_in_forward_at_455 2025-09-07T08:08:07.5212032Z context_layer = context_layer.contiguous().view(batch_size, from_seq_length, -1) 2025-09-07T08:08:07.5212265Z 2025-09-07T08:08:07.7235351Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:07.7236012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T08:08:07.7236596Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:08:07.7237081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T08:08:07.7237512Z hidden_states = self.dense(hidden_states) 2025-09-07T08:08:07.7237655Z 2025-09-07T08:08:07.9783784Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:07.9784406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:08:07.9784923Z layer_output = apply_chunking_to_forward( 2025-09-07T08:08:07.9785421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:07.9786377Z return forward_fn(*input_tensors) 2025-09-07T08:08:07.9786922Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:08:07.9787406Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:08:07.9787872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T08:08:07.9788317Z hidden_states = self.dense(hidden_states) 2025-09-07T08:08:07.9788462Z 2025-09-07T08:08:07.9788575Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:07.9789076Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:08:07.9789558Z layer_output = apply_chunking_to_forward( 2025-09-07T08:08:07.9789966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:07.9790527Z return forward_fn(*input_tensors) 2025-09-07T08:08:07.9790963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T08:08:07.9791433Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:08:07.9791882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T08:08:07.9792322Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:08:07.9792711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:08:07.9793179Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:08:07.9793421Z 2025-09-07T08:08:07.9793528Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:07.9794030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T08:08:07.9794511Z layer_output = apply_chunking_to_forward( 2025-09-07T08:08:07.9794915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:07.9795316Z return forward_fn(*input_tensors) 2025-09-07T08:08:07.9795732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1513, in feed_forward_chunk 2025-09-07T08:08:07.9796215Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:08:07.9796674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1415, in forward 2025-09-07T08:08:07.9797083Z hidden_states = self.dense(hidden_states) 2025-09-07T08:08:07.9797220Z 2025-09-07T08:08:08.1999464Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:08.2000006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T08:08:08.2000427Z self.query(hidden_states) 2025-09-07T08:08:08.2000546Z 2025-09-07T08:08:08.2000660Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:08.2001092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 445, in forward 2025-09-07T08:08:08.2001491Z self.key(hidden_states) 2025-09-07T08:08:08.2001611Z 2025-09-07T08:08:08.2001710Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:08.2002140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 450, in forward 2025-09-07T08:08:08.2002543Z self.value(hidden_states) 2025-09-07T08:08:08.2002999Z 2025-09-07T08:08:10.8497127Z W0907 08:08:10.848857 49872 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] torch._dynamo hit config.recompile_limit (8) 2025-09-07T08:08:10.8498494Z W0907 08:08:10.848857 49872 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] function: 'forward' (/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py:417) 2025-09-07T08:08:10.8500289Z W0907 08:08:10.848857 49872 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] last reason: 10/7: self.seed == 7 # context_layer, attention_probs = self.bigbird_block_sparse_attention( # transformers/models/big_bird/modeling_big_bird.py:455 in forward (HINT: torch.compile considers integer attributes of the nn.Module to be static. If you are observing recompilation, you might want to make this integer dynamic using torch._dynamo.config.allow_unspec_int_on_nn_module = True, or convert this integer into a tensor.) 2025-09-07T08:08:10.8502085Z W0907 08:08:10.848857 49872 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] To log all recompilation reasons, use TORCH_LOGS="recompiles". 2025-09-07T08:08:10.8502901Z W0907 08:08:10.848857 49872 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] To diagnose recompilation issues, see https://pytorch.org/docs/main/torch.compiler_troubleshooting.html 2025-09-07T08:08:11.5237863Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:11.5238474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2010, in torch_dynamo_resume_in_forward_at_1991 2025-09-07T08:08:11.5239144Z pooler_output = self.activation(self.pooler(sequence_output[:, 0, :])) if (self.pooler is not None) else None 2025-09-07T08:08:11.5239429Z 2025-09-07T08:08:11.5239538Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:11.5240049Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2010, in torch_dynamo_resume_in_forward_at_1991 2025-09-07T08:08:11.5240701Z pooler_output = self.activation(self.pooler(sequence_output[:, 0, :])) if (self.pooler is not None) else None 2025-09-07T08:08:11.5240967Z 2025-09-07T08:08:11.7438611Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:11.7439220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2318, in torch_dynamo_resume_in_forward_at_2303 2025-09-07T08:08:11.7439738Z prediction_scores = self.cls(sequence_output) 2025-09-07T08:08:11.7440188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1682, in forward 2025-09-07T08:08:11.7440645Z prediction_scores = self.predictions(sequence_output) 2025-09-07T08:08:11.7441081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1670, in forward 2025-09-07T08:08:11.7441519Z hidden_states = self.transform(hidden_states) 2025-09-07T08:08:11.7441951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1645, in forward 2025-09-07T08:08:11.7442354Z hidden_states = self.dense(hidden_states) 2025-09-07T08:08:11.7442492Z 2025-09-07T08:08:11.7442603Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:11.7443100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2318, in torch_dynamo_resume_in_forward_at_2303 2025-09-07T08:08:11.7443578Z prediction_scores = self.cls(sequence_output) 2025-09-07T08:08:11.7443987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1682, in forward 2025-09-07T08:08:11.7444420Z prediction_scores = self.predictions(sequence_output) 2025-09-07T08:08:11.7444844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1670, in forward 2025-09-07T08:08:11.7445775Z hidden_states = self.transform(hidden_states) 2025-09-07T08:08:11.7446219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1645, in forward 2025-09-07T08:08:11.7446632Z hidden_states = self.dense(hidden_states) 2025-09-07T08:08:11.7446772Z 2025-09-07T08:08:11.7446875Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:11.7447373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2318, in torch_dynamo_resume_in_forward_at_2303 2025-09-07T08:08:11.7447862Z prediction_scores = self.cls(sequence_output) 2025-09-07T08:08:11.7448288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1682, in forward 2025-09-07T08:08:11.7448730Z prediction_scores = self.predictions(sequence_output) 2025-09-07T08:08:11.7449320Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1671, in forward 2025-09-07T08:08:11.7449740Z hidden_states = self.decoder(hidden_states) 2025-09-07T08:08:11.7449880Z 2025-09-07T08:08:13.9147809Z pass 2025-09-07T08:08:13.9148178Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:08:16.8338168Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:08:16.8339286Z import pynvml # type: ignore[import] 2025-09-07T08:08:18.8354191Z 2025-09-07T08:08:21.4277487Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:08:21.4277871Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:08:21.4415902Z cpu eval hf_DistilBert 2025-09-07T08:08:22.0345195Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:08:22.2056557Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:08:22.4210670Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:08:27.2102713Z cudagraph partition due to non gpu ops 2025-09-07T08:08:27.2103010Z cudagraph partition due to non gpu ops 2025-09-07T08:08:27.2103332Z cudagraph partition due to non gpu ops 2025-09-07T08:08:27.2103531Z cudagraph partition due to non gpu ops 2025-09-07T08:08:27.2103724Z cudagraph partition due to non gpu ops 2025-09-07T08:08:27.2103908Z cudagraph partition due to non gpu ops 2025-09-07T08:08:27.2104138Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2104540Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2104940Z return mod(**inputs) 2025-09-07T08:08:27.2105381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2105841Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2106286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2106712Z return self.transformer( 2025-09-07T08:08:27.2107124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2107531Z layer_outputs = layer_module( 2025-09-07T08:08:27.2107885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2108250Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2109015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2109537Z sa_output = self.attention( 2025-09-07T08:08:27.2109933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 390, in forward 2025-09-07T08:08:27.2110397Z q = shape(self.q_lin(query)) # (bs, n_heads, q_length, dim_per_head) 2025-09-07T08:08:27.2110582Z 2025-09-07T08:08:27.2110686Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2111065Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2111430Z return mod(**inputs) 2025-09-07T08:08:27.2111812Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2112229Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2112786Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2113205Z return self.transformer( 2025-09-07T08:08:27.2113588Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2114010Z layer_outputs = layer_module( 2025-09-07T08:08:27.2114363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2114729Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2115146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2115558Z sa_output = self.attention( 2025-09-07T08:08:27.2115951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 391, in forward 2025-09-07T08:08:27.2116408Z k = shape(self.k_lin(key)) # (bs, n_heads, k_length, dim_per_head) 2025-09-07T08:08:27.2116585Z 2025-09-07T08:08:27.2116694Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2117045Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2117356Z return mod(**inputs) 2025-09-07T08:08:27.2117735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2118144Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2118539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2118937Z return self.transformer( 2025-09-07T08:08:27.2119331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2119736Z layer_outputs = layer_module( 2025-09-07T08:08:27.2120079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2120434Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2120833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2121238Z sa_output = self.attention( 2025-09-07T08:08:27.2121621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 392, in forward 2025-09-07T08:08:27.2122076Z v = shape(self.v_lin(value)) # (bs, n_heads, k_length, dim_per_head) 2025-09-07T08:08:27.2122250Z 2025-09-07T08:08:27.2122333Z cudagraph partition due to non gpu ops 2025-09-07T08:08:27.2122684Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2123031Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2123435Z return mod(**inputs) 2025-09-07T08:08:27.2123821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2124235Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2124627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2125032Z return self.transformer( 2025-09-07T08:08:27.2125421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2125829Z layer_outputs = layer_module( 2025-09-07T08:08:27.2126155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2126508Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2126989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2127405Z sa_output = self.attention( 2025-09-07T08:08:27.2127790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 402, in forward 2025-09-07T08:08:27.2128247Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:27.2128435Z 2025-09-07T08:08:27.2128530Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2128869Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2129186Z return mod(**inputs) 2025-09-07T08:08:27.2129565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2129968Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2130367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2130786Z return self.transformer( 2025-09-07T08:08:27.2131167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2131581Z layer_outputs = layer_module( 2025-09-07T08:08:27.2131909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2132261Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2132666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2133071Z sa_output = self.attention( 2025-09-07T08:08:27.2133457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 412, in forward 2025-09-07T08:08:27.2133870Z attn_output = self.out_lin(attn_output) 2025-09-07T08:08:27.2134011Z 2025-09-07T08:08:27.2134107Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2134448Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2134760Z return mod(**inputs) 2025-09-07T08:08:27.2135134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2135535Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2135929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2136332Z return self.transformer( 2025-09-07T08:08:27.2136713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2137118Z layer_outputs = layer_module( 2025-09-07T08:08:27.2137577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2137977Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2138386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2138828Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2139260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2139802Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2140331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2140725Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2141223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 431, in ff_chunk 2025-09-07T08:08:27.2141634Z x = self.lin1(input) 2025-09-07T08:08:27.2141739Z 2025-09-07T08:08:27.2141837Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2142191Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2142507Z return mod(**inputs) 2025-09-07T08:08:27.2142886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2143292Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2143692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2144112Z return self.transformer( 2025-09-07T08:08:27.2144506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2144928Z layer_outputs = layer_module( 2025-09-07T08:08:27.2145258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2145617Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2146031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2146476Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2146909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2147447Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2147972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2148378Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2148790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 432, in ff_chunk 2025-09-07T08:08:27.2149201Z x = self.activation(x) 2025-09-07T08:08:27.2149523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:08:27.2149858Z return self.act(input) 2025-09-07T08:08:27.2149959Z 2025-09-07T08:08:27.2150068Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2150420Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2150737Z return mod(**inputs) 2025-09-07T08:08:27.2151121Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2151584Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2152033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2152438Z return self.transformer( 2025-09-07T08:08:27.2152830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2153240Z layer_outputs = layer_module( 2025-09-07T08:08:27.2153577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2153941Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2154342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2154783Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2155220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2155862Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2156384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2156772Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2157189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 433, in ff_chunk 2025-09-07T08:08:27.2157595Z x = self.lin2(x) 2025-09-07T08:08:27.2157690Z 2025-09-07T08:08:27.2157793Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2158143Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2158465Z return mod(**inputs) 2025-09-07T08:08:27.2158868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2159290Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2159704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2160110Z return self.transformer( 2025-09-07T08:08:27.2160503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2160912Z layer_outputs = layer_module( 2025-09-07T08:08:27.2161259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2161620Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2162027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2162440Z sa_output = self.attention( 2025-09-07T08:08:27.2162838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 390, in forward 2025-09-07T08:08:27.2163301Z q = shape(self.q_lin(query)) # (bs, n_heads, q_length, dim_per_head) 2025-09-07T08:08:27.2163477Z 2025-09-07T08:08:27.2163581Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2163930Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2164250Z return mod(**inputs) 2025-09-07T08:08:27.2164633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2165044Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2165440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2165906Z return self.transformer( 2025-09-07T08:08:27.2166339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2166753Z layer_outputs = layer_module( 2025-09-07T08:08:27.2167096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2167449Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2167862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2168269Z sa_output = self.attention( 2025-09-07T08:08:27.2168663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 391, in forward 2025-09-07T08:08:27.2169121Z k = shape(self.k_lin(key)) # (bs, n_heads, k_length, dim_per_head) 2025-09-07T08:08:27.2169297Z 2025-09-07T08:08:27.2169395Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2169828Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2170149Z return mod(**inputs) 2025-09-07T08:08:27.2170533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2170946Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2171338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2171743Z return self.transformer( 2025-09-07T08:08:27.2172133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2172539Z layer_outputs = layer_module( 2025-09-07T08:08:27.2172871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2173230Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2173640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2174047Z sa_output = self.attention( 2025-09-07T08:08:27.2174436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 392, in forward 2025-09-07T08:08:27.2174884Z v = shape(self.v_lin(value)) # (bs, n_heads, k_length, dim_per_head) 2025-09-07T08:08:27.2175062Z 2025-09-07T08:08:27.2175141Z cudagraph partition due to non gpu ops 2025-09-07T08:08:27.2175370Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2175715Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2176026Z return mod(**inputs) 2025-09-07T08:08:27.2176408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2176821Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2177217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2177687Z return self.transformer( 2025-09-07T08:08:27.2178080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2178491Z layer_outputs = layer_module( 2025-09-07T08:08:27.2178827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2179184Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2179595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2180057Z sa_output = self.attention( 2025-09-07T08:08:27.2180491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 402, in forward 2025-09-07T08:08:27.2180957Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:27.2181139Z 2025-09-07T08:08:27.2181242Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2181589Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2181904Z return mod(**inputs) 2025-09-07T08:08:27.2182280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2182687Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2183089Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2183492Z return self.transformer( 2025-09-07T08:08:27.2183946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2184361Z layer_outputs = layer_module( 2025-09-07T08:08:27.2184697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2185050Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2185452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2185859Z sa_output = self.attention( 2025-09-07T08:08:27.2186246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 412, in forward 2025-09-07T08:08:27.2186665Z attn_output = self.out_lin(attn_output) 2025-09-07T08:08:27.2186796Z 2025-09-07T08:08:27.2186903Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2187245Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2187573Z return mod(**inputs) 2025-09-07T08:08:27.2187973Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2188395Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2188798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2189213Z return self.transformer( 2025-09-07T08:08:27.2189621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2190037Z layer_outputs = layer_module( 2025-09-07T08:08:27.2190382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2190744Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2191165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2191613Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2192058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2192598Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2193114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2193513Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2193926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 431, in ff_chunk 2025-09-07T08:08:27.2194391Z x = self.lin1(input) 2025-09-07T08:08:27.2194532Z 2025-09-07T08:08:27.2194644Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2194997Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2195323Z return mod(**inputs) 2025-09-07T08:08:27.2195706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2196120Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2196515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2196921Z return self.transformer( 2025-09-07T08:08:27.2197315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2197731Z layer_outputs = layer_module( 2025-09-07T08:08:27.2198147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2198509Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2198920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2199363Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2199805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2200339Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2200852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2201246Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2201659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 432, in ff_chunk 2025-09-07T08:08:27.2202075Z x = self.activation(x) 2025-09-07T08:08:27.2202389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:08:27.2202741Z return self.act(input) 2025-09-07T08:08:27.2202850Z 2025-09-07T08:08:27.2202947Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2203470Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2203789Z return mod(**inputs) 2025-09-07T08:08:27.2204165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2204579Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2204977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2205393Z return self.transformer( 2025-09-07T08:08:27.2205783Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2206187Z layer_outputs = layer_module( 2025-09-07T08:08:27.2206525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2206878Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2207286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2207720Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2208158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2208795Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2209400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2218838Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2219382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 433, in ff_chunk 2025-09-07T08:08:27.2219892Z x = self.lin2(x) 2025-09-07T08:08:27.2220005Z 2025-09-07T08:08:27.2220113Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2220491Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2220824Z return mod(**inputs) 2025-09-07T08:08:27.2221241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2221694Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2222294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2222726Z return self.transformer( 2025-09-07T08:08:27.2223128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2223542Z layer_outputs = layer_module( 2025-09-07T08:08:27.2223885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2224251Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2224674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2225093Z sa_output = self.attention( 2025-09-07T08:08:27.2225486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 390, in forward 2025-09-07T08:08:27.2225962Z q = shape(self.q_lin(query)) # (bs, n_heads, q_length, dim_per_head) 2025-09-07T08:08:27.2226147Z 2025-09-07T08:08:27.2226251Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2226621Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2226954Z return mod(**inputs) 2025-09-07T08:08:27.2227338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2227754Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2228161Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2228581Z return self.transformer( 2025-09-07T08:08:27.2228983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2229392Z layer_outputs = layer_module( 2025-09-07T08:08:27.2229740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2230109Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2230535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2230939Z sa_output = self.attention( 2025-09-07T08:08:27.2231336Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 391, in forward 2025-09-07T08:08:27.2231793Z k = shape(self.k_lin(key)) # (bs, n_heads, k_length, dim_per_head) 2025-09-07T08:08:27.2231969Z 2025-09-07T08:08:27.2232079Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2232436Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2234305Z return mod(**inputs) 2025-09-07T08:08:27.2234710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2235133Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2235543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2235963Z return self.transformer( 2025-09-07T08:08:27.2236347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2236754Z layer_outputs = layer_module( 2025-09-07T08:08:27.2237101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2237468Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2237952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2238372Z sa_output = self.attention( 2025-09-07T08:08:27.2238765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 392, in forward 2025-09-07T08:08:27.2239226Z v = shape(self.v_lin(value)) # (bs, n_heads, k_length, dim_per_head) 2025-09-07T08:08:27.2239405Z 2025-09-07T08:08:27.2239491Z cudagraph partition due to non gpu ops 2025-09-07T08:08:27.2239718Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2240074Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2240399Z return mod(**inputs) 2025-09-07T08:08:27.2240790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2241200Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2241602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2242011Z return self.transformer( 2025-09-07T08:08:27.2242419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2242830Z layer_outputs = layer_module( 2025-09-07T08:08:27.2243165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2243526Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2243940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2244353Z sa_output = self.attention( 2025-09-07T08:08:27.2244740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 402, in forward 2025-09-07T08:08:27.2245205Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:27.2245392Z 2025-09-07T08:08:27.2245489Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2245837Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2246157Z return mod(**inputs) 2025-09-07T08:08:27.2246536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2246937Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2247333Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2247747Z return self.transformer( 2025-09-07T08:08:27.2248136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2248629Z layer_outputs = layer_module( 2025-09-07T08:08:27.2248984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2249359Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2249812Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2250250Z sa_output = self.attention( 2025-09-07T08:08:27.2250651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 412, in forward 2025-09-07T08:08:27.2251075Z attn_output = self.out_lin(attn_output) 2025-09-07T08:08:27.2251216Z 2025-09-07T08:08:27.2251313Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2251668Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2251990Z return mod(**inputs) 2025-09-07T08:08:27.2252443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2252866Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2253288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2253711Z return self.transformer( 2025-09-07T08:08:27.2254105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2254522Z layer_outputs = layer_module( 2025-09-07T08:08:27.2254875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2255245Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2255661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2256114Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2256567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2257121Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2257703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2258111Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2258522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 431, in ff_chunk 2025-09-07T08:08:27.2258935Z x = self.lin1(input) 2025-09-07T08:08:27.2259044Z 2025-09-07T08:08:27.2259148Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2259519Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2259843Z return mod(**inputs) 2025-09-07T08:08:27.2260220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2260635Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2261041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2261452Z return self.transformer( 2025-09-07T08:08:27.2261833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2262240Z layer_outputs = layer_module( 2025-09-07T08:08:27.2262584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2263006Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2263464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2263926Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2264365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2264903Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2265428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2265825Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2266271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 432, in ff_chunk 2025-09-07T08:08:27.2266692Z x = self.activation(x) 2025-09-07T08:08:27.2267087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:08:27.2267433Z return self.act(input) 2025-09-07T08:08:27.2267533Z 2025-09-07T08:08:27.2267639Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2267986Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2268310Z return mod(**inputs) 2025-09-07T08:08:27.2268698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2269115Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2269510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2269920Z return self.transformer( 2025-09-07T08:08:27.2270322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2270734Z layer_outputs = layer_module( 2025-09-07T08:08:27.2271076Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2271428Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2271841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2272286Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2272730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2273267Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2273780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2274183Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2274596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 433, in ff_chunk 2025-09-07T08:08:27.2275002Z x = self.lin2(x) 2025-09-07T08:08:27.2275095Z 2025-09-07T08:08:27.2275200Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2275546Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2275871Z return mod(**inputs) 2025-09-07T08:08:27.2276254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2276667Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2277058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2277571Z return self.transformer( 2025-09-07T08:08:27.2277967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2278380Z layer_outputs = layer_module( 2025-09-07T08:08:27.2278719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2279071Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2279481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2279888Z sa_output = self.attention( 2025-09-07T08:08:27.2280285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 390, in forward 2025-09-07T08:08:27.2280754Z q = shape(self.q_lin(query)) # (bs, n_heads, q_length, dim_per_head) 2025-09-07T08:08:27.2280933Z 2025-09-07T08:08:27.2281096Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2281452Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2281768Z return mod(**inputs) 2025-09-07T08:08:27.2282147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2282553Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2282949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2283356Z return self.transformer( 2025-09-07T08:08:27.2283743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2284159Z layer_outputs = layer_module( 2025-09-07T08:08:27.2284503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2284862Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2285272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2285690Z sa_output = self.attention( 2025-09-07T08:08:27.2286075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 391, in forward 2025-09-07T08:08:27.2286531Z k = shape(self.k_lin(key)) # (bs, n_heads, k_length, dim_per_head) 2025-09-07T08:08:27.2286705Z 2025-09-07T08:08:27.2286800Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2287149Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2287464Z return mod(**inputs) 2025-09-07T08:08:27.2287842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2288254Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2288653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2289055Z return self.transformer( 2025-09-07T08:08:27.2289432Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2289835Z layer_outputs = layer_module( 2025-09-07T08:08:27.2290164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2290516Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2290919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2291366Z sa_output = self.attention( 2025-09-07T08:08:27.2291787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 392, in forward 2025-09-07T08:08:27.2292235Z v = shape(self.v_lin(value)) # (bs, n_heads, k_length, dim_per_head) 2025-09-07T08:08:27.2292406Z 2025-09-07T08:08:27.2292488Z cudagraph partition due to non gpu ops 2025-09-07T08:08:27.2292711Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2293048Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2293360Z return mod(**inputs) 2025-09-07T08:08:27.2293736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2294146Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2294535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2295013Z return self.transformer( 2025-09-07T08:08:27.2295398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2295803Z layer_outputs = layer_module( 2025-09-07T08:08:27.2296130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2296482Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2296883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2297286Z sa_output = self.attention( 2025-09-07T08:08:27.2297732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 402, in forward 2025-09-07T08:08:27.2298199Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:27.2298384Z 2025-09-07T08:08:27.2298481Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2298834Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2299146Z return mod(**inputs) 2025-09-07T08:08:27.2299519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2299921Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2300311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2300710Z return self.transformer( 2025-09-07T08:08:27.2301092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2301491Z layer_outputs = layer_module( 2025-09-07T08:08:27.2301818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2302169Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2302565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2302967Z sa_output = self.attention( 2025-09-07T08:08:27.2303518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 412, in forward 2025-09-07T08:08:27.2303932Z attn_output = self.out_lin(attn_output) 2025-09-07T08:08:27.2304067Z 2025-09-07T08:08:27.2304163Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2304503Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2304816Z return mod(**inputs) 2025-09-07T08:08:27.2305183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2305726Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2306138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2306542Z return self.transformer( 2025-09-07T08:08:27.2306925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2307318Z layer_outputs = layer_module( 2025-09-07T08:08:27.2307648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2307997Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2308402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2308845Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2309380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2309925Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2310452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2310854Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2311266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 431, in ff_chunk 2025-09-07T08:08:27.2311671Z x = self.lin1(input) 2025-09-07T08:08:27.2311774Z 2025-09-07T08:08:27.2311871Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2312220Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2312535Z return mod(**inputs) 2025-09-07T08:08:27.2312905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2313313Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2313713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2314118Z return self.transformer( 2025-09-07T08:08:27.2314503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2314903Z layer_outputs = layer_module( 2025-09-07T08:08:27.2315234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2315582Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2315995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2316430Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2316858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2317393Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2317911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2318304Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2318715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 432, in ff_chunk 2025-09-07T08:08:27.2319123Z x = self.activation(x) 2025-09-07T08:08:27.2319440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:08:27.2319869Z return self.act(input) 2025-09-07T08:08:27.2319980Z 2025-09-07T08:08:27.2320089Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2320441Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2320759Z return mod(**inputs) 2025-09-07T08:08:27.2321139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2321549Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2321951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2322349Z return self.transformer( 2025-09-07T08:08:27.2322735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2323144Z layer_outputs = layer_module( 2025-09-07T08:08:27.2323569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2323936Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2324344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2324788Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2325224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2325759Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2326272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2326662Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2327073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 433, in ff_chunk 2025-09-07T08:08:27.2327476Z x = self.lin2(x) 2025-09-07T08:08:27.2327569Z 2025-09-07T08:08:27.2327671Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2328015Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2328334Z return mod(**inputs) 2025-09-07T08:08:27.2328712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2329117Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2329511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2329909Z return self.transformer( 2025-09-07T08:08:27.2330296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2330705Z layer_outputs = layer_module( 2025-09-07T08:08:27.2331041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2331396Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2331793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2332202Z sa_output = self.attention( 2025-09-07T08:08:27.2332593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 390, in forward 2025-09-07T08:08:27.2333055Z q = shape(self.q_lin(query)) # (bs, n_heads, q_length, dim_per_head) 2025-09-07T08:08:27.2333228Z 2025-09-07T08:08:27.2333323Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2333720Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2334071Z return mod(**inputs) 2025-09-07T08:08:27.2334450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2334856Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2335245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2335647Z return self.transformer( 2025-09-07T08:08:27.2336027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2336425Z layer_outputs = layer_module( 2025-09-07T08:08:27.2336751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2337098Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2337614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2338021Z sa_output = self.attention( 2025-09-07T08:08:27.2338402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 391, in forward 2025-09-07T08:08:27.2338842Z k = shape(self.k_lin(key)) # (bs, n_heads, k_length, dim_per_head) 2025-09-07T08:08:27.2339016Z 2025-09-07T08:08:27.2339108Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2339443Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2339745Z return mod(**inputs) 2025-09-07T08:08:27.2340111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2340509Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2340908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2341311Z return self.transformer( 2025-09-07T08:08:27.2341690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2342093Z layer_outputs = layer_module( 2025-09-07T08:08:27.2342416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2342766Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2343169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2343572Z sa_output = self.attention( 2025-09-07T08:08:27.2343947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 392, in forward 2025-09-07T08:08:27.2344401Z v = shape(self.v_lin(value)) # (bs, n_heads, k_length, dim_per_head) 2025-09-07T08:08:27.2344576Z 2025-09-07T08:08:27.2344650Z cudagraph partition due to non gpu ops 2025-09-07T08:08:27.2344873Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2345213Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2345520Z return mod(**inputs) 2025-09-07T08:08:27.2345896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2346301Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2346695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2347096Z return self.transformer( 2025-09-07T08:08:27.2347531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2347982Z layer_outputs = layer_module( 2025-09-07T08:08:27.2348327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2348683Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2349093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2349501Z sa_output = self.attention( 2025-09-07T08:08:27.2349895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 402, in forward 2025-09-07T08:08:27.2350359Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:27.2350542Z 2025-09-07T08:08:27.2350651Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2351070Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2351403Z return mod(**inputs) 2025-09-07T08:08:27.2351790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2352206Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2352613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2353015Z return self.transformer( 2025-09-07T08:08:27.2353405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2353817Z layer_outputs = layer_module( 2025-09-07T08:08:27.2354154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2354506Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2354932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2355344Z sa_output = self.attention( 2025-09-07T08:08:27.2355742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 412, in forward 2025-09-07T08:08:27.2356160Z attn_output = self.out_lin(attn_output) 2025-09-07T08:08:27.2356293Z 2025-09-07T08:08:27.2356396Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2356862Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2357191Z return mod(**inputs) 2025-09-07T08:08:27.2357579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2357997Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2358396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2358809Z return self.transformer( 2025-09-07T08:08:27.2359201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2359610Z layer_outputs = layer_module( 2025-09-07T08:08:27.2359946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2360305Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2360712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2361160Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2361600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2362230Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2362756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2363159Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2363584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 431, in ff_chunk 2025-09-07T08:08:27.2363993Z x = self.lin1(input) 2025-09-07T08:08:27.2364093Z 2025-09-07T08:08:27.2364193Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2364544Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2364862Z return mod(**inputs) 2025-09-07T08:08:27.2365243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2365733Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2366131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2366539Z return self.transformer( 2025-09-07T08:08:27.2366923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2367334Z layer_outputs = layer_module( 2025-09-07T08:08:27.2367663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2368022Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2368433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2368878Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2369324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2369855Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2370373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2370768Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2371178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 432, in ff_chunk 2025-09-07T08:08:27.2371593Z x = self.activation(x) 2025-09-07T08:08:27.2371907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:08:27.2372243Z return self.act(input) 2025-09-07T08:08:27.2372350Z 2025-09-07T08:08:27.2372447Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2372802Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2373122Z return mod(**inputs) 2025-09-07T08:08:27.2373492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2373906Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2374309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2374716Z return self.transformer( 2025-09-07T08:08:27.2375099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2375508Z layer_outputs = layer_module( 2025-09-07T08:08:27.2375845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2376289Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2376704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2377144Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2377625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2378166Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2378687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2379085Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2379493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 433, in ff_chunk 2025-09-07T08:08:27.2380002Z x = self.lin2(x) 2025-09-07T08:08:27.2380106Z 2025-09-07T08:08:27.2380205Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2380559Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2380883Z return mod(**inputs) 2025-09-07T08:08:27.2381284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2381704Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2382104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2382513Z return self.transformer( 2025-09-07T08:08:27.2382904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2383321Z layer_outputs = layer_module( 2025-09-07T08:08:27.2383668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2384026Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2384438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2384840Z sa_output = self.attention( 2025-09-07T08:08:27.2385235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 390, in forward 2025-09-07T08:08:27.2385692Z q = shape(self.q_lin(query)) # (bs, n_heads, q_length, dim_per_head) 2025-09-07T08:08:27.2385866Z 2025-09-07T08:08:27.2385970Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2386316Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2386629Z return mod(**inputs) 2025-09-07T08:08:27.2387013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2387425Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2387827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2388229Z return self.transformer( 2025-09-07T08:08:27.2388620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2389045Z layer_outputs = layer_module( 2025-09-07T08:08:27.2389393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2389751Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2390159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2390662Z sa_output = self.attention( 2025-09-07T08:08:27.2391057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 391, in forward 2025-09-07T08:08:27.2391519Z k = shape(self.k_lin(key)) # (bs, n_heads, k_length, dim_per_head) 2025-09-07T08:08:27.2391695Z 2025-09-07T08:08:27.2391797Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2392141Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2392456Z return mod(**inputs) 2025-09-07T08:08:27.2392837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2393248Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2393645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2394136Z return self.transformer( 2025-09-07T08:08:27.2394524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2394934Z layer_outputs = layer_module( 2025-09-07T08:08:27.2395269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2395620Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2396025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2396432Z sa_output = self.attention( 2025-09-07T08:08:27.2396819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 392, in forward 2025-09-07T08:08:27.2397279Z v = shape(self.v_lin(value)) # (bs, n_heads, k_length, dim_per_head) 2025-09-07T08:08:27.2397455Z 2025-09-07T08:08:27.2397539Z cudagraph partition due to non gpu ops 2025-09-07T08:08:27.2397766Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2398113Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2398428Z return mod(**inputs) 2025-09-07T08:08:27.2398798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2399209Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2399608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2400012Z return self.transformer( 2025-09-07T08:08:27.2400399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2400803Z layer_outputs = layer_module( 2025-09-07T08:08:27.2401143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2401498Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2401911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2402318Z sa_output = self.attention( 2025-09-07T08:08:27.2402699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 402, in forward 2025-09-07T08:08:27.2403299Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:27.2403486Z 2025-09-07T08:08:27.2403581Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2403928Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2404316Z return mod(**inputs) 2025-09-07T08:08:27.2404759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2405174Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2405575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2405979Z return self.transformer( 2025-09-07T08:08:27.2406358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2406770Z layer_outputs = layer_module( 2025-09-07T08:08:27.2407102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2407458Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2407862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 476, in forward 2025-09-07T08:08:27.2408566Z sa_output = self.attention( 2025-09-07T08:08:27.2408959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 412, in forward 2025-09-07T08:08:27.2409378Z attn_output = self.out_lin(attn_output) 2025-09-07T08:08:27.2409512Z 2025-09-07T08:08:27.2409611Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2409959Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2410272Z return mod(**inputs) 2025-09-07T08:08:27.2410652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2411062Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2411454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2411868Z return self.transformer( 2025-09-07T08:08:27.2412253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2412656Z layer_outputs = layer_module( 2025-09-07T08:08:27.2412984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2413334Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2413727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2414176Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2414618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2415153Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2415679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2416065Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2416474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 431, in ff_chunk 2025-09-07T08:08:27.2416879Z x = self.lin1(input) 2025-09-07T08:08:27.2416977Z 2025-09-07T08:08:27.2417074Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2417470Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2417790Z return mod(**inputs) 2025-09-07T08:08:27.2418160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2418627Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2419073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2419482Z return self.transformer( 2025-09-07T08:08:27.2419866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2420271Z layer_outputs = layer_module( 2025-09-07T08:08:27.2420603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2420960Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2421361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2421796Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2422232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2422834Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2423356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2423749Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2424151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 432, in ff_chunk 2025-09-07T08:08:27.2424561Z x = self.activation(x) 2025-09-07T08:08:27.2424877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:08:27.2425206Z return self.act(input) 2025-09-07T08:08:27.2425307Z 2025-09-07T08:08:27.2425402Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2425755Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2426068Z return mod(**inputs) 2025-09-07T08:08:27.2426442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 826, in forward 2025-09-07T08:08:27.2426856Z dlbrt_output = self.distilbert( 2025-09-07T08:08:27.2427261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 736, in forward 2025-09-07T08:08:27.2427665Z return self.transformer( 2025-09-07T08:08:27.2428055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 541, in forward 2025-09-07T08:08:27.2428462Z layer_outputs = layer_module( 2025-09-07T08:08:27.2428793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:27.2429151Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:27.2429568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 494, in forward 2025-09-07T08:08:27.2430015Z ffn_output = self.ffn(sa_output) # (bs, seq_length, dim) 2025-09-07T08:08:27.2430447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 428, in forward 2025-09-07T08:08:27.2430983Z return apply_chunking_to_forward(self.ff_chunk, self.chunk_size_feed_forward, self.seq_len_dim, input) 2025-09-07T08:08:27.2431495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:08:27.2431891Z return forward_fn(*input_tensors) 2025-09-07T08:08:27.2432301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 433, in ff_chunk 2025-09-07T08:08:27.2432702Z x = self.lin2(x) 2025-09-07T08:08:27.2432847Z 2025-09-07T08:08:27.2432944Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2433357Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2433677Z return mod(**inputs) 2025-09-07T08:08:27.2434054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 836, in forward 2025-09-07T08:08:27.2434557Z prediction_logits = self.vocab_transform(hidden_states) # (bs, seq_length, dim) 2025-09-07T08:08:27.2434779Z 2025-09-07T08:08:27.2434871Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:27.2435218Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:27.2435530Z return mod(**inputs) 2025-09-07T08:08:27.2435913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/distilbert/modeling_distilbert.py", line 839, in forward 2025-09-07T08:08:27.2436495Z prediction_logits = self.vocab_projector(prediction_logits) # (bs, seq_length, vocab_size) 2025-09-07T08:08:27.2436745Z 2025-09-07T08:08:33.5649233Z pass 2025-09-07T08:08:33.5649632Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:08:36.0415806Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:08:36.0416783Z import pynvml # type: ignore[import] 2025-09-07T08:08:38.0275312Z 2025-09-07T08:08:43.6220596Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:08:43.6221086Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:08:43.6402373Z cpu eval hf_GPT2 2025-09-07T08:08:44.1817677Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:08:44.3453836Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:08:44.4996005Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:08:51.8923736Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8924041Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8924252Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8924446Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8924643Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8924835Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8925030Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8925223Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8925417Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8925624Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8925857Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8926064Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8926327Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.8926713Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.8927052Z return mod(**inputs) 2025-09-07T08:08:51.8927464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.8927878Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.8928289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.8928679Z outputs = block( 2025-09-07T08:08:51.8929036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.8929443Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.8930209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.8930717Z return func(*args, **kwargs) 2025-09-07T08:08:51.8931097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.8931558Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.8931963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.8932336Z return func(*args, **kwargs) 2025-09-07T08:08:51.8932710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:08:51.8933210Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:08:51.8933681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.8934285Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.8934469Z 2025-09-07T08:08:51.8934551Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8934753Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8934946Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8935136Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8935358Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.8935728Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.8936054Z return mod(**inputs) 2025-09-07T08:08:51.8936452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.8936864Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.8937248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.8937689Z outputs = block( 2025-09-07T08:08:51.8938049Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.8938424Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.8938803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.8939174Z return func(*args, **kwargs) 2025-09-07T08:08:51.8939537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.8939940Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.8940319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.8940677Z return func(*args, **kwargs) 2025-09-07T08:08:51.8941032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.8941436Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.8941891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:08:51.8942382Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:51.8942568Z 2025-09-07T08:08:51.8942670Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.8943029Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.8943350Z return mod(**inputs) 2025-09-07T08:08:51.8943720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.8944124Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.8944513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.8944975Z outputs = block( 2025-09-07T08:08:51.8945297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.8945655Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.8946020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.8946394Z return func(*args, **kwargs) 2025-09-07T08:08:51.8946755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.8947146Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.8947520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.8947889Z return func(*args, **kwargs) 2025-09-07T08:08:51.8948314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.8948730Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.8949178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:08:51.8949632Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:08:51.8949798Z 2025-09-07T08:08:51.8949898Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.8950257Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.8950580Z return mod(**inputs) 2025-09-07T08:08:51.8950931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.8951322Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.8951707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.8952077Z outputs = block( 2025-09-07T08:08:51.8952402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.8952758Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.8953119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.8953484Z return func(*args, **kwargs) 2025-09-07T08:08:51.8953844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.8954229Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.8954606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.8955008Z return func(*args, **kwargs) 2025-09-07T08:08:51.8955363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:08:51.8955760Z attn_output = self.c_proj(attn_output) 2025-09-07T08:08:51.8956110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.8956497Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.8956670Z 2025-09-07T08:08:51.8956766Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.8957140Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.8957458Z return mod(**inputs) 2025-09-07T08:08:51.8957808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.8958191Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.8958619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.8959029Z outputs = block( 2025-09-07T08:08:51.8959365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.8959711Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.8960077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.8960442Z return func(*args, **kwargs) 2025-09-07T08:08:51.8960799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.8961208Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.8961602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:08:51.8961983Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:08:51.8962423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.8962847Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.8963015Z 2025-09-07T08:08:51.8963119Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.8963460Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.8974354Z return mod(**inputs) 2025-09-07T08:08:51.8974775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.8975211Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.8975630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.8976031Z outputs = block( 2025-09-07T08:08:51.8976376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.8976749Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.8977134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.8977509Z return func(*args, **kwargs) 2025-09-07T08:08:51.8977989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.8978415Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.8978827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:08:51.8979217Z hidden_states = self.act(hidden_states) 2025-09-07T08:08:51.8979562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:08:51.8980024Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:08:51.8980271Z 2025-09-07T08:08:51.8980375Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.8980741Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.8981060Z return mod(**inputs) 2025-09-07T08:08:51.8981426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.8981831Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.8982213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.8982581Z outputs = block( 2025-09-07T08:08:51.8982892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.8983328Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.8983755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.8984129Z return func(*args, **kwargs) 2025-09-07T08:08:51.8984493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.8984892Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.8985288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:08:51.8985681Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:08:51.8986045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.8986446Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.8986628Z 2025-09-07T08:08:51.8986728Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.8987153Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.8987482Z return mod(**inputs) 2025-09-07T08:08:51.8987836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.8988227Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.8988606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.8988974Z outputs = block( 2025-09-07T08:08:51.8989289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.8989650Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.8990019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.8990393Z return func(*args, **kwargs) 2025-09-07T08:08:51.8990759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 442, in forward 2025-09-07T08:08:51.8991169Z hidden_states = residual + feed_forward_hidden_states 2025-09-07T08:08:51.8991325Z 2025-09-07T08:08:51.8991424Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.8991771Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.8992085Z return mod(**inputs) 2025-09-07T08:08:51.8992437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.8992827Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.8993211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.8993590Z outputs = block( 2025-09-07T08:08:51.8993915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.8994281Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.8994654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.8995027Z return func(*args, **kwargs) 2025-09-07T08:08:51.8995397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.8995797Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.8996180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.8996545Z return func(*args, **kwargs) 2025-09-07T08:08:51.8996902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:08:51.8997450Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:08:51.8997970Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.8998371Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.8998543Z 2025-09-07T08:08:51.8998621Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8998824Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8999017Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8999207Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.8999430Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.8999791Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9000117Z return mod(**inputs) 2025-09-07T08:08:51.9000476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9000943Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9001327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9001703Z outputs = block( 2025-09-07T08:08:51.9002028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9002390Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9002763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9003329Z return func(*args, **kwargs) 2025-09-07T08:08:51.9003695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9004093Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9004487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9004850Z return func(*args, **kwargs) 2025-09-07T08:08:51.9005208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9005614Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9006066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:08:51.9006551Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:51.9006734Z 2025-09-07T08:08:51.9006835Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9007185Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9007500Z return mod(**inputs) 2025-09-07T08:08:51.9007858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9008244Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9008626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9008994Z outputs = block( 2025-09-07T08:08:51.9009308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9009663Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9010028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9010399Z return func(*args, **kwargs) 2025-09-07T08:08:51.9010760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9011251Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9011685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9012060Z return func(*args, **kwargs) 2025-09-07T08:08:51.9012426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9012829Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9013275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:08:51.9013751Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:08:51.9013923Z 2025-09-07T08:08:51.9014021Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9014378Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9014702Z return mod(**inputs) 2025-09-07T08:08:51.9015146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9015547Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9015938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9016307Z outputs = block( 2025-09-07T08:08:51.9016625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9016973Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9017351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9017785Z return func(*args, **kwargs) 2025-09-07T08:08:51.9018149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9018544Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9018917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9019289Z return func(*args, **kwargs) 2025-09-07T08:08:51.9019640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:08:51.9020027Z attn_output = self.c_proj(attn_output) 2025-09-07T08:08:51.9020386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9020780Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9020958Z 2025-09-07T08:08:51.9021062Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9021417Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9021745Z return mod(**inputs) 2025-09-07T08:08:51.9022098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9022498Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9022885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9023258Z outputs = block( 2025-09-07T08:08:51.9023574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9023925Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9024301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9024670Z return func(*args, **kwargs) 2025-09-07T08:08:51.9025031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9025498Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9025932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:08:51.9026319Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:08:51.9026670Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9027060Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9027228Z 2025-09-07T08:08:51.9027328Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9027680Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9027997Z return mod(**inputs) 2025-09-07T08:08:51.9028349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9028745Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9029196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9029569Z outputs = block( 2025-09-07T08:08:51.9029887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9030248Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9030613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9030982Z return func(*args, **kwargs) 2025-09-07T08:08:51.9031339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9031742Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9032141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:08:51.9032521Z hidden_states = self.act(hidden_states) 2025-09-07T08:08:51.9032857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:08:51.9033310Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:08:51.9033542Z 2025-09-07T08:08:51.9033647Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9033995Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9034313Z return mod(**inputs) 2025-09-07T08:08:51.9034663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9035057Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9035437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9035808Z outputs = block( 2025-09-07T08:08:51.9036122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9036475Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9036843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9037217Z return func(*args, **kwargs) 2025-09-07T08:08:51.9037577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9037973Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9038372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:08:51.9038754Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:08:51.9039160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9039591Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9039760Z 2025-09-07T08:08:51.9039862Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9040201Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9040524Z return mod(**inputs) 2025-09-07T08:08:51.9040877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9041271Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9041659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9042023Z outputs = block( 2025-09-07T08:08:51.9042336Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9042757Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9043130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9043494Z return func(*args, **kwargs) 2025-09-07T08:08:51.9043858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9044249Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9044631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9044995Z return func(*args, **kwargs) 2025-09-07T08:08:51.9045340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:08:51.9045833Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:08:51.9046301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9046707Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9046872Z 2025-09-07T08:08:51.9046964Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9047162Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9047359Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9047555Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9047778Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9048127Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9048446Z return mod(**inputs) 2025-09-07T08:08:51.9048799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9049188Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9049570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9049925Z outputs = block( 2025-09-07T08:08:51.9050231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9050582Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9050954Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9051315Z return func(*args, **kwargs) 2025-09-07T08:08:51.9051672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9052062Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9052439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9052881Z return func(*args, **kwargs) 2025-09-07T08:08:51.9053233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9053628Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9054079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:08:51.9054560Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:51.9054738Z 2025-09-07T08:08:51.9054838Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9055180Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9055504Z return mod(**inputs) 2025-09-07T08:08:51.9055853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9056241Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9056672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9057043Z outputs = block( 2025-09-07T08:08:51.9057357Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9057814Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9058184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9058545Z return func(*args, **kwargs) 2025-09-07T08:08:51.9058902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9059288Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9059662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9060028Z return func(*args, **kwargs) 2025-09-07T08:08:51.9060380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9060770Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9061205Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:08:51.9061654Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:08:51.9061811Z 2025-09-07T08:08:51.9061901Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9062235Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9062540Z return mod(**inputs) 2025-09-07T08:08:51.9062881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9063267Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9063635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9063996Z outputs = block( 2025-09-07T08:08:51.9064301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9064643Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9064998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9065356Z return func(*args, **kwargs) 2025-09-07T08:08:51.9065700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9066076Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9066510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9066959Z return func(*args, **kwargs) 2025-09-07T08:08:51.9067313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:08:51.9067694Z attn_output = self.c_proj(attn_output) 2025-09-07T08:08:51.9068036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9068415Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9068579Z 2025-09-07T08:08:51.9068671Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9069009Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9069318Z return mod(**inputs) 2025-09-07T08:08:51.9069668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9070124Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9070514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9070881Z outputs = block( 2025-09-07T08:08:51.9071192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9071553Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9071940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9072297Z return func(*args, **kwargs) 2025-09-07T08:08:51.9072671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9073075Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9073472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:08:51.9073845Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:08:51.9074213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9074589Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9074753Z 2025-09-07T08:08:51.9074849Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9075185Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9075490Z return mod(**inputs) 2025-09-07T08:08:51.9075828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9076208Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9076587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9076948Z outputs = block( 2025-09-07T08:08:51.9077268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9077638Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9078001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9078361Z return func(*args, **kwargs) 2025-09-07T08:08:51.9078706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9079150Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9079570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:08:51.9079958Z hidden_states = self.act(hidden_states) 2025-09-07T08:08:51.9080349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:08:51.9080845Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:08:51.9081077Z 2025-09-07T08:08:51.9081167Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9081522Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9081840Z return mod(**inputs) 2025-09-07T08:08:51.9082179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9082560Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9082924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9083344Z outputs = block( 2025-09-07T08:08:51.9083711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9084063Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9084422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9084779Z return func(*args, **kwargs) 2025-09-07T08:08:51.9085133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9085522Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9085915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:08:51.9086293Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:08:51.9086642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9087025Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9087192Z 2025-09-07T08:08:51.9087285Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9087623Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9087927Z return mod(**inputs) 2025-09-07T08:08:51.9088265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9088649Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9089013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9089385Z outputs = block( 2025-09-07T08:08:51.9089717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9090066Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9090433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9090805Z return func(*args, **kwargs) 2025-09-07T08:08:51.9091152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 442, in forward 2025-09-07T08:08:51.9091551Z hidden_states = residual + feed_forward_hidden_states 2025-09-07T08:08:51.9091706Z 2025-09-07T08:08:51.9091799Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9092156Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9092482Z return mod(**inputs) 2025-09-07T08:08:51.9092844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9093221Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9093635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9094068Z outputs = block( 2025-09-07T08:08:51.9094377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9094740Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9095114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9095476Z return func(*args, **kwargs) 2025-09-07T08:08:51.9095819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9096195Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9096563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9096938Z return func(*args, **kwargs) 2025-09-07T08:08:51.9097351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:08:51.9097923Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:08:51.9098395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9098791Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9098955Z 2025-09-07T08:08:51.9099178Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9099371Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9099555Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9099738Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9099976Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9100323Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9100644Z return mod(**inputs) 2025-09-07T08:08:51.9100998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9101392Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9101762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9102131Z outputs = block( 2025-09-07T08:08:51.9102436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9102793Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9103275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9103639Z return func(*args, **kwargs) 2025-09-07T08:08:51.9104000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9104394Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9104769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9105136Z return func(*args, **kwargs) 2025-09-07T08:08:51.9105481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9105877Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9106322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:08:51.9106800Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:51.9106978Z 2025-09-07T08:08:51.9107074Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9107492Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9107870Z return mod(**inputs) 2025-09-07T08:08:51.9108236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9108632Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9109019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9109383Z outputs = block( 2025-09-07T08:08:51.9109691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9110047Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9110419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9110781Z return func(*args, **kwargs) 2025-09-07T08:08:51.9112390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9112811Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9113197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9113561Z return func(*args, **kwargs) 2025-09-07T08:08:51.9113923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9114326Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9114772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:08:51.9115229Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:08:51.9115386Z 2025-09-07T08:08:51.9115485Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9115849Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9116169Z return mod(**inputs) 2025-09-07T08:08:51.9116525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9116915Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9117298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9117663Z outputs = block( 2025-09-07T08:08:51.9117977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9118337Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9118706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9119077Z return func(*args, **kwargs) 2025-09-07T08:08:51.9119442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9119836Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9120218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9120581Z return func(*args, **kwargs) 2025-09-07T08:08:51.9120935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:08:51.9121320Z attn_output = self.c_proj(attn_output) 2025-09-07T08:08:51.9121677Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9122067Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9122245Z 2025-09-07T08:08:51.9122342Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9122745Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9123123Z return mod(**inputs) 2025-09-07T08:08:51.9123476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9123860Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9124237Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9124603Z outputs = block( 2025-09-07T08:08:51.9124918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9125281Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9125646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9126018Z return func(*args, **kwargs) 2025-09-07T08:08:51.9126441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9126856Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9127253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:08:51.9127633Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:08:51.9127978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9128386Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9128552Z 2025-09-07T08:08:51.9128655Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9128995Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9129312Z return mod(**inputs) 2025-09-07T08:08:51.9129664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9130060Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9130433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9130796Z outputs = block( 2025-09-07T08:08:51.9131103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9131459Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9131830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9132190Z return func(*args, **kwargs) 2025-09-07T08:08:51.9132542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9132944Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9133348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:08:51.9133730Z hidden_states = self.act(hidden_states) 2025-09-07T08:08:51.9134062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:08:51.9134512Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:08:51.9134747Z 2025-09-07T08:08:51.9134839Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9135206Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9135525Z return mod(**inputs) 2025-09-07T08:08:51.9135867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9136319Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9136738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9137113Z outputs = block( 2025-09-07T08:08:51.9137415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9137855Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9138235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9138608Z return func(*args, **kwargs) 2025-09-07T08:08:51.9138964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9139359Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9139754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:08:51.9140215Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:08:51.9140575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9140984Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9141157Z 2025-09-07T08:08:51.9141252Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9141595Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9141908Z return mod(**inputs) 2025-09-07T08:08:51.9142256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9142639Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9143013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9143381Z outputs = block( 2025-09-07T08:08:51.9143693Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9144045Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9144405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9144772Z return func(*args, **kwargs) 2025-09-07T08:08:51.9145125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9145511Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9145883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9146248Z return func(*args, **kwargs) 2025-09-07T08:08:51.9146599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:08:51.9147096Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:08:51.9147556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9147939Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9148112Z 2025-09-07T08:08:51.9148190Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9148390Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9148584Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9148776Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9148984Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9149334Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9149651Z return mod(**inputs) 2025-09-07T08:08:51.9150053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9150476Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9150851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9151217Z outputs = block( 2025-09-07T08:08:51.9151529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9151885Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9152250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9152612Z return func(*args, **kwargs) 2025-09-07T08:08:51.9152966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9153353Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9153785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9154154Z return func(*args, **kwargs) 2025-09-07T08:08:51.9154511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9154913Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9155362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:08:51.9155836Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:51.9156020Z 2025-09-07T08:08:51.9156112Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9156458Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9156777Z return mod(**inputs) 2025-09-07T08:08:51.9157131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9157512Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9157890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9158254Z outputs = block( 2025-09-07T08:08:51.9158563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9158909Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9159272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9159634Z return func(*args, **kwargs) 2025-09-07T08:08:51.9159986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9160370Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9160738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9161099Z return func(*args, **kwargs) 2025-09-07T08:08:51.9161446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9161837Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9162270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:08:51.9162727Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:08:51.9162889Z 2025-09-07T08:08:51.9162981Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9163325Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9163690Z return mod(**inputs) 2025-09-07T08:08:51.9164073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9164459Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9164837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9165202Z outputs = block( 2025-09-07T08:08:51.9165515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9165860Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9166228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9166594Z return func(*args, **kwargs) 2025-09-07T08:08:51.9166948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9167398Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9167777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9168148Z return func(*args, **kwargs) 2025-09-07T08:08:51.9168502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:08:51.9168884Z attn_output = self.c_proj(attn_output) 2025-09-07T08:08:51.9169224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9169610Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9169786Z 2025-09-07T08:08:51.9169879Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9170225Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9170544Z return mod(**inputs) 2025-09-07T08:08:51.9170890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9171278Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9171656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9172023Z outputs = block( 2025-09-07T08:08:51.9172322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9172677Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9173042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9173409Z return func(*args, **kwargs) 2025-09-07T08:08:51.9173762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9174177Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9174577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:08:51.9174958Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:08:51.9175314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9175703Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9175869Z 2025-09-07T08:08:51.9175962Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9176306Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9176618Z return mod(**inputs) 2025-09-07T08:08:51.9176962Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9177425Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9177854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9178224Z outputs = block( 2025-09-07T08:08:51.9178529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9178880Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9179238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9179601Z return func(*args, **kwargs) 2025-09-07T08:08:51.9179956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9180353Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9180745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:08:51.9181212Z hidden_states = self.act(hidden_states) 2025-09-07T08:08:51.9181558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:08:51.9182008Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:08:51.9182238Z 2025-09-07T08:08:51.9182341Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9182677Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9182989Z return mod(**inputs) 2025-09-07T08:08:51.9183339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9183731Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9184108Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9184470Z outputs = block( 2025-09-07T08:08:51.9184779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9185135Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9185501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9185866Z return func(*args, **kwargs) 2025-09-07T08:08:51.9186218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9186615Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9187011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:08:51.9187399Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:08:51.9187747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9188138Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9188310Z 2025-09-07T08:08:51.9188402Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9188751Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9189065Z return mod(**inputs) 2025-09-07T08:08:51.9189405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9189794Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9190168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9190536Z outputs = block( 2025-09-07T08:08:51.9190882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9191283Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9191655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9192025Z return func(*args, **kwargs) 2025-09-07T08:08:51.9192386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 442, in forward 2025-09-07T08:08:51.9192801Z hidden_states = residual + feed_forward_hidden_states 2025-09-07T08:08:51.9192966Z 2025-09-07T08:08:51.9193060Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9193401Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9193715Z return mod(**inputs) 2025-09-07T08:08:51.9194058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9194508Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9194890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9195257Z outputs = block( 2025-09-07T08:08:51.9195564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9195910Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9196288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9196665Z return func(*args, **kwargs) 2025-09-07T08:08:51.9197048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9197457Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9197862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9198238Z return func(*args, **kwargs) 2025-09-07T08:08:51.9198611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:08:51.9199110Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:08:51.9199576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9199971Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9200147Z 2025-09-07T08:08:51.9200224Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9200431Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9200627Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9200816Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9201049Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9201418Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9201745Z return mod(**inputs) 2025-09-07T08:08:51.9202101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9202502Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9202898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9203410Z outputs = block( 2025-09-07T08:08:51.9203737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9204093Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9204470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9204917Z return func(*args, **kwargs) 2025-09-07T08:08:51.9205342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9205735Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9206113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9206482Z return func(*args, **kwargs) 2025-09-07T08:08:51.9206840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9207238Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9207680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:08:51.9208164Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:51.9208353Z 2025-09-07T08:08:51.9208453Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9208894Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9209223Z return mod(**inputs) 2025-09-07T08:08:51.9209575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9209972Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9210356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9210724Z outputs = block( 2025-09-07T08:08:51.9211038Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9211391Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9211763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9212140Z return func(*args, **kwargs) 2025-09-07T08:08:51.9212502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9212882Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9213271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9213647Z return func(*args, **kwargs) 2025-09-07T08:08:51.9214021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9214431Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9214875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:08:51.9215337Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:08:51.9215507Z 2025-09-07T08:08:51.9215610Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9215970Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9216292Z return mod(**inputs) 2025-09-07T08:08:51.9216643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9217039Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9217425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9217861Z outputs = block( 2025-09-07T08:08:51.9218176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9218537Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9218965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9219385Z return func(*args, **kwargs) 2025-09-07T08:08:51.9219752Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9220143Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9220531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9220900Z return func(*args, **kwargs) 2025-09-07T08:08:51.9221258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:08:51.9221650Z attn_output = self.c_proj(attn_output) 2025-09-07T08:08:51.9221999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9222402Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9222585Z 2025-09-07T08:08:51.9222750Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9223114Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9223430Z return mod(**inputs) 2025-09-07T08:08:51.9223791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9224186Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9224573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9224940Z outputs = block( 2025-09-07T08:08:51.9225247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9225605Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9225976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9226350Z return func(*args, **kwargs) 2025-09-07T08:08:51.9226702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9227104Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9227503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:08:51.9227883Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:08:51.9228232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9228613Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9228789Z 2025-09-07T08:08:51.9228883Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9229234Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9229559Z return mod(**inputs) 2025-09-07T08:08:51.9229915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9230304Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9230679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9231049Z outputs = block( 2025-09-07T08:08:51.9231358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9231703Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9232071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9232437Z return func(*args, **kwargs) 2025-09-07T08:08:51.9232845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9233284Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9233677Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:08:51.9234059Z hidden_states = self.act(hidden_states) 2025-09-07T08:08:51.9234399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:08:51.9234851Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:08:51.9235085Z 2025-09-07T08:08:51.9235199Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9235543Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9235875Z return mod(**inputs) 2025-09-07T08:08:51.9236311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9236712Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9237090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9237482Z outputs = block( 2025-09-07T08:08:51.9237802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9238182Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9238556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9238933Z return func(*args, **kwargs) 2025-09-07T08:08:51.9239292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9239694Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9240117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:08:51.9240499Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:08:51.9240853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9241244Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9241414Z 2025-09-07T08:08:51.9241515Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9241867Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9242176Z return mod(**inputs) 2025-09-07T08:08:51.9242515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9242905Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9250281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9250679Z outputs = block( 2025-09-07T08:08:51.9251027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9251399Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9251775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9252149Z return func(*args, **kwargs) 2025-09-07T08:08:51.9252528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9252961Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9253365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9253816Z return func(*args, **kwargs) 2025-09-07T08:08:51.9254231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:08:51.9254730Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:08:51.9255191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9255582Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9255762Z 2025-09-07T08:08:51.9255841Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9256041Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9256233Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9256421Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9256637Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9256997Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9257382Z return mod(**inputs) 2025-09-07T08:08:51.9257837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9258231Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9258616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9258984Z outputs = block( 2025-09-07T08:08:51.9259300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9259655Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9260019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9260381Z return func(*args, **kwargs) 2025-09-07T08:08:51.9260743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9261129Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9261497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9261855Z return func(*args, **kwargs) 2025-09-07T08:08:51.9262203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9262596Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9263037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:08:51.9263510Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:51.9263695Z 2025-09-07T08:08:51.9263797Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9264149Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9264487Z return mod(**inputs) 2025-09-07T08:08:51.9264834Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9265217Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9265593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9265952Z outputs = block( 2025-09-07T08:08:51.9266274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9266620Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9266988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9267402Z return func(*args, **kwargs) 2025-09-07T08:08:51.9267765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9268185Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9268548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9268910Z return func(*args, **kwargs) 2025-09-07T08:08:51.9269256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9269643Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9270082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:08:51.9270534Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:08:51.9270697Z 2025-09-07T08:08:51.9270793Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9271198Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9271518Z return mod(**inputs) 2025-09-07T08:08:51.9271865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9272248Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9272623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9272990Z outputs = block( 2025-09-07T08:08:51.9273296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9273645Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9274014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9274378Z return func(*args, **kwargs) 2025-09-07T08:08:51.9274737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9275121Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9275490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9275851Z return func(*args, **kwargs) 2025-09-07T08:08:51.9276204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:08:51.9276587Z attn_output = self.c_proj(attn_output) 2025-09-07T08:08:51.9276932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9277322Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9277495Z 2025-09-07T08:08:51.9277594Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9277944Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9278254Z return mod(**inputs) 2025-09-07T08:08:51.9278598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9278996Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9279368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9279731Z outputs = block( 2025-09-07T08:08:51.9280033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9280383Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9280739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9281144Z return func(*args, **kwargs) 2025-09-07T08:08:51.9281536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9281931Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9282331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:08:51.9282713Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:08:51.9283059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9283444Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9283609Z 2025-09-07T08:08:51.9283705Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9284054Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9284368Z return mod(**inputs) 2025-09-07T08:08:51.9284782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9285173Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9285556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9285912Z outputs = block( 2025-09-07T08:08:51.9286211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9286553Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9286907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9287263Z return func(*args, **kwargs) 2025-09-07T08:08:51.9287606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9288001Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9288390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:08:51.9288762Z hidden_states = self.act(hidden_states) 2025-09-07T08:08:51.9289092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:08:51.9289534Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:08:51.9289773Z 2025-09-07T08:08:51.9289869Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9290213Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9290525Z return mod(**inputs) 2025-09-07T08:08:51.9290901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9291300Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9291677Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9292041Z outputs = block( 2025-09-07T08:08:51.9292352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9292701Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9293068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9293435Z return func(*args, **kwargs) 2025-09-07T08:08:51.9293782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9294177Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9294662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:08:51.9295305Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:08:51.9295648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9296041Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9296209Z 2025-09-07T08:08:51.9296302Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9296649Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9296958Z return mod(**inputs) 2025-09-07T08:08:51.9297294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9297741Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9298111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9298544Z outputs = block( 2025-09-07T08:08:51.9298880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9299239Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9299612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9299976Z return func(*args, **kwargs) 2025-09-07T08:08:51.9300334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 442, in forward 2025-09-07T08:08:51.9300732Z hidden_states = residual + feed_forward_hidden_states 2025-09-07T08:08:51.9300894Z 2025-09-07T08:08:51.9300987Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9301323Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9301638Z return mod(**inputs) 2025-09-07T08:08:51.9301979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9302355Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9302721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9303227Z outputs = block( 2025-09-07T08:08:51.9303532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9303880Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9304245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9304612Z return func(*args, **kwargs) 2025-09-07T08:08:51.9304977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9305372Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9305747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9306105Z return func(*args, **kwargs) 2025-09-07T08:08:51.9306454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:08:51.9306947Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:08:51.9307407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9307797Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9307968Z 2025-09-07T08:08:51.9308043Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9308246Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9308508Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9308744Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9308962Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9309310Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9309624Z return mod(**inputs) 2025-09-07T08:08:51.9309972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9310356Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9310729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9311087Z outputs = block( 2025-09-07T08:08:51.9311395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9311747Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9312218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9312582Z return func(*args, **kwargs) 2025-09-07T08:08:51.9312934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9313312Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9313687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9314047Z return func(*args, **kwargs) 2025-09-07T08:08:51.9314409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9314802Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9315241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:08:51.9315724Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:51.9315904Z 2025-09-07T08:08:51.9315998Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9316342Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9316401Z return mod(**inputs) 2025-09-07T08:08:51.9316653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9316726Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9316967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9317025Z outputs = block( 2025-09-07T08:08:51.9317243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9317316Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9317556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9317618Z return func(*args, **kwargs) 2025-09-07T08:08:51.9317859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9317934Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9318170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9318229Z return func(*args, **kwargs) 2025-09-07T08:08:51.9318468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9318553Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9318839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:08:51.9319011Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:08:51.9319015Z 2025-09-07T08:08:51.9319106Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9319294Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9319352Z return mod(**inputs) 2025-09-07T08:08:51.9319601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9319675Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9319916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9319972Z outputs = block( 2025-09-07T08:08:51.9320188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9320259Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9320556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9320617Z return func(*args, **kwargs) 2025-09-07T08:08:51.9320861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9320935Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9321169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9321231Z return func(*args, **kwargs) 2025-09-07T08:08:51.9321469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:08:51.9321542Z attn_output = self.c_proj(attn_output) 2025-09-07T08:08:51.9321758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9321869Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9321873Z 2025-09-07T08:08:51.9321964Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9322150Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9322213Z return mod(**inputs) 2025-09-07T08:08:51.9322459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9322533Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9322772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9322826Z outputs = block( 2025-09-07T08:08:51.9323042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9323115Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9323351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9323412Z return func(*args, **kwargs) 2025-09-07T08:08:51.9323654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9323753Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9323993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:08:51.9324066Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:08:51.9324277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9324385Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9324433Z 2025-09-07T08:08:51.9324557Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9324742Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9324802Z return mod(**inputs) 2025-09-07T08:08:51.9325046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9325119Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9325359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9325413Z outputs = block( 2025-09-07T08:08:51.9325625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9325694Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9325929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9326051Z return func(*args, **kwargs) 2025-09-07T08:08:51.9326299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9326390Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9326632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:08:51.9326706Z hidden_states = self.act(hidden_states) 2025-09-07T08:08:51.9326915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:08:51.9327093Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:08:51.9327097Z 2025-09-07T08:08:51.9327188Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9327378Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9327441Z return mod(**inputs) 2025-09-07T08:08:51.9327688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9327761Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9328000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9328057Z outputs = block( 2025-09-07T08:08:51.9328271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9328338Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9328578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9328638Z return func(*args, **kwargs) 2025-09-07T08:08:51.9328882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9328973Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9329211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:08:51.9329290Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:08:51.9329502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9329608Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9329611Z 2025-09-07T08:08:51.9329699Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9329882Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9329941Z return mod(**inputs) 2025-09-07T08:08:51.9330240Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9330345Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9330587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9330643Z outputs = block( 2025-09-07T08:08:51.9330857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9330923Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9331162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9331222Z return func(*args, **kwargs) 2025-09-07T08:08:51.9331463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9331542Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9331837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9331900Z return func(*args, **kwargs) 2025-09-07T08:08:51.9332140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:08:51.9332324Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:08:51.9332534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9332636Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9332639Z 2025-09-07T08:08:51.9332714Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9332781Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9332850Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9332919Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9333013Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9333198Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9333255Z return mod(**inputs) 2025-09-07T08:08:51.9333505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9333575Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9333815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9333869Z outputs = block( 2025-09-07T08:08:51.9334081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9334150Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9334385Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9334451Z return func(*args, **kwargs) 2025-09-07T08:08:51.9334691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9334765Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9334998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9335058Z return func(*args, **kwargs) 2025-09-07T08:08:51.9335297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9335384Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9335679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:08:51.9335842Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:51.9335880Z 2025-09-07T08:08:51.9335976Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9336166Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9336224Z return mod(**inputs) 2025-09-07T08:08:51.9336477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9336552Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9336794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9336852Z outputs = block( 2025-09-07T08:08:51.9337064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9337136Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9337441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9337504Z return func(*args, **kwargs) 2025-09-07T08:08:51.9337815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9337897Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9338143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9338205Z return func(*args, **kwargs) 2025-09-07T08:08:51.9338449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9338540Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9338833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:08:51.9338940Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:08:51.9338947Z 2025-09-07T08:08:51.9339044Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9339239Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9339297Z return mod(**inputs) 2025-09-07T08:08:51.9339545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9339622Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9339863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9339923Z outputs = block( 2025-09-07T08:08:51.9340140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9340213Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9340456Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9340518Z return func(*args, **kwargs) 2025-09-07T08:08:51.9340762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9340838Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9341072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9341132Z return func(*args, **kwargs) 2025-09-07T08:08:51.9341370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:08:51.9341444Z attn_output = self.c_proj(attn_output) 2025-09-07T08:08:51.9341658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9341820Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9341859Z 2025-09-07T08:08:51.9341953Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9342140Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9342199Z return mod(**inputs) 2025-09-07T08:08:51.9342446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9342521Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9342761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9342817Z outputs = block( 2025-09-07T08:08:51.9343031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9343099Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9343403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9343465Z return func(*args, **kwargs) 2025-09-07T08:08:51.9343707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9343802Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9344042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:08:51.9344115Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:08:51.9344323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9344427Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9344431Z 2025-09-07T08:08:51.9344520Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9344712Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9344770Z return mod(**inputs) 2025-09-07T08:08:51.9345014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9345089Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9345327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9345383Z outputs = block( 2025-09-07T08:08:51.9345595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9345662Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9345896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9345956Z return func(*args, **kwargs) 2025-09-07T08:08:51.9346202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9346290Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9346529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:08:51.9346600Z hidden_states = self.act(hidden_states) 2025-09-07T08:08:51.9346805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:08:51.9346979Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:08:51.9346982Z 2025-09-07T08:08:51.9347069Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9347254Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9347350Z return mod(**inputs) 2025-09-07T08:08:51.9347658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9347748Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9347991Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9348046Z outputs = block( 2025-09-07T08:08:51.9348257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9348324Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9348561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9348621Z return func(*args, **kwargs) 2025-09-07T08:08:51.9348864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9349032Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9349275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:08:51.9349355Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:08:51.9349564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9349673Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9349676Z 2025-09-07T08:08:51.9349767Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9349952Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9350009Z return mod(**inputs) 2025-09-07T08:08:51.9350255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9350332Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9350575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9350636Z outputs = block( 2025-09-07T08:08:51.9350846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9350912Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9351147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9351206Z return func(*args, **kwargs) 2025-09-07T08:08:51.9351447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 442, in forward 2025-09-07T08:08:51.9351546Z hidden_states = residual + feed_forward_hidden_states 2025-09-07T08:08:51.9351550Z 2025-09-07T08:08:51.9351641Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9351829Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9351888Z return mod(**inputs) 2025-09-07T08:08:51.9352137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9352207Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9352448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9352504Z outputs = block( 2025-09-07T08:08:51.9352713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9352786Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9353022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9353122Z return func(*args, **kwargs) 2025-09-07T08:08:51.9353401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9353481Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9353721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9353784Z return func(*args, **kwargs) 2025-09-07T08:08:51.9354025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:08:51.9354207Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:08:51.9354419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9354522Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9354527Z 2025-09-07T08:08:51.9354602Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9354733Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9354802Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9354870Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9354960Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9355145Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9355205Z return mod(**inputs) 2025-09-07T08:08:51.9355450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9355525Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9355765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9355818Z outputs = block( 2025-09-07T08:08:51.9356033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9356105Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9356341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9356400Z return func(*args, **kwargs) 2025-09-07T08:08:51.9356638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9356714Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9356947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9357007Z return func(*args, **kwargs) 2025-09-07T08:08:51.9357245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9357336Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9357630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:08:51.9357748Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:51.9357751Z 2025-09-07T08:08:51.9357843Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9358026Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9358083Z return mod(**inputs) 2025-09-07T08:08:51.9358326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9358397Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9358638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9358733Z outputs = block( 2025-09-07T08:08:51.9358949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9359048Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9359285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9359343Z return func(*args, **kwargs) 2025-09-07T08:08:51.9359581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9359656Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9359891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9359953Z return func(*args, **kwargs) 2025-09-07T08:08:51.9360192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9360274Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9360624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:08:51.9360723Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:08:51.9360726Z 2025-09-07T08:08:51.9360819Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9361004Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9361062Z return mod(**inputs) 2025-09-07T08:08:51.9361307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9361376Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9361621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9361676Z outputs = block( 2025-09-07T08:08:51.9361893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9361962Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9362201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9362265Z return func(*args, **kwargs) 2025-09-07T08:08:51.9362509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9362587Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9362823Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9362887Z return func(*args, **kwargs) 2025-09-07T08:08:51.9363130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:08:51.9363208Z attn_output = self.c_proj(attn_output) 2025-09-07T08:08:51.9363424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9363531Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9363534Z 2025-09-07T08:08:51.9363627Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9363815Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9363873Z return mod(**inputs) 2025-09-07T08:08:51.9364123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9364201Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9364442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9364531Z outputs = block( 2025-09-07T08:08:51.9364778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9364848Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9365083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9365145Z return func(*args, **kwargs) 2025-09-07T08:08:51.9365384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9365479Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9365720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:08:51.9365789Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:08:51.9366003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9366163Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9366166Z 2025-09-07T08:08:51.9366258Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9366442Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9366499Z return mod(**inputs) 2025-09-07T08:08:51.9366747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9366817Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9367062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9367116Z outputs = block( 2025-09-07T08:08:51.9367328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9367399Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9367636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9367713Z return func(*args, **kwargs) 2025-09-07T08:08:51.9367951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9368040Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9368282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:08:51.9368351Z hidden_states = self.act(hidden_states) 2025-09-07T08:08:51.9368559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:08:51.9368728Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:08:51.9368733Z 2025-09-07T08:08:51.9368826Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9369008Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9369064Z return mod(**inputs) 2025-09-07T08:08:51.9369314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9369385Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9369626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9369679Z outputs = block( 2025-09-07T08:08:51.9369890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9369957Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9370190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9370316Z return func(*args, **kwargs) 2025-09-07T08:08:51.9370556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9370644Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9370889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:08:51.9370969Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:08:51.9371182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9371289Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9371292Z 2025-09-07T08:08:51.9371385Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9371571Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9371702Z return mod(**inputs) 2025-09-07T08:08:51.9371953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9372028Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9372276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9372330Z outputs = block( 2025-09-07T08:08:51.9372544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9372618Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9372855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9372919Z return func(*args, **kwargs) 2025-09-07T08:08:51.9373159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9373243Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9373476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9373535Z return func(*args, **kwargs) 2025-09-07T08:08:51.9373776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:08:51.9373953Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:08:51.9374166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9374268Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9374271Z 2025-09-07T08:08:51.9374345Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9374415Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9374486Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9374555Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9374647Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9374833Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9374892Z return mod(**inputs) 2025-09-07T08:08:51.9375137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9375212Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9375450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9375506Z outputs = block( 2025-09-07T08:08:51.9375719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9375819Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9376089Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9376150Z return func(*args, **kwargs) 2025-09-07T08:08:51.9376392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9376466Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9376698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9376759Z return func(*args, **kwargs) 2025-09-07T08:08:51.9376997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9377084Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9377449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:08:51.9377617Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:51.9377622Z 2025-09-07T08:08:51.9377713Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9377905Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9377964Z return mod(**inputs) 2025-09-07T08:08:51.9378223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9378296Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9378538Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9378591Z outputs = block( 2025-09-07T08:08:51.9378806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9378877Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9379114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9379172Z return func(*args, **kwargs) 2025-09-07T08:08:51.9379414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9379489Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9379720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9379783Z return func(*args, **kwargs) 2025-09-07T08:08:51.9380035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9380129Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9380429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:08:51.9380534Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:08:51.9380537Z 2025-09-07T08:08:51.9380635Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9380834Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9380896Z return mod(**inputs) 2025-09-07T08:08:51.9381148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9381224Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9381474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9381530Z outputs = block( 2025-09-07T08:08:51.9381753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9382062Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9382306Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9382373Z return func(*args, **kwargs) 2025-09-07T08:08:51.9382614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9382696Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9382931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9382995Z return func(*args, **kwargs) 2025-09-07T08:08:51.9383234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:08:51.9383307Z attn_output = self.c_proj(attn_output) 2025-09-07T08:08:51.9383591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9383707Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9383710Z 2025-09-07T08:08:51.9383808Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9384003Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9384061Z return mod(**inputs) 2025-09-07T08:08:51.9384321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9384396Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9384641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9384695Z outputs = block( 2025-09-07T08:08:51.9384914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9384988Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9385227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9385288Z return func(*args, **kwargs) 2025-09-07T08:08:51.9385527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9385625Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9385865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:08:51.9385934Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:08:51.9386149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9386254Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9386259Z 2025-09-07T08:08:51.9386354Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9386541Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9386597Z return mod(**inputs) 2025-09-07T08:08:51.9386844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9386914Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9387155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9387210Z outputs = block( 2025-09-07T08:08:51.9387423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9387489Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9387778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9387876Z return func(*args, **kwargs) 2025-09-07T08:08:51.9388116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9388210Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9388450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:08:51.9388520Z hidden_states = self.act(hidden_states) 2025-09-07T08:08:51.9388731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:08:51.9388906Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:08:51.9388910Z 2025-09-07T08:08:51.9389003Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9389252Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9389312Z return mod(**inputs) 2025-09-07T08:08:51.9389560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9389632Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9389876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9389930Z outputs = block( 2025-09-07T08:08:51.9390144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9390211Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9390446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9390510Z return func(*args, **kwargs) 2025-09-07T08:08:51.9390758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9390849Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9391088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:08:51.9391164Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:08:51.9391377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9391481Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9391484Z 2025-09-07T08:08:51.9391577Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9391763Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9391825Z return mod(**inputs) 2025-09-07T08:08:51.9392076Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9392154Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9392400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9392457Z outputs = block( 2025-09-07T08:08:51.9392675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9392743Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9392977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9393040Z return func(*args, **kwargs) 2025-09-07T08:08:51.9393279Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 442, in forward 2025-09-07T08:08:51.9393435Z hidden_states = residual + feed_forward_hidden_states 2025-09-07T08:08:51.9393490Z 2025-09-07T08:08:51.9393584Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9393776Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9393833Z return mod(**inputs) 2025-09-07T08:08:51.9394078Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9394152Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9394390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9394446Z outputs = block( 2025-09-07T08:08:51.9394661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9394727Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9395068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9395131Z return func(*args, **kwargs) 2025-09-07T08:08:51.9395375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9395454Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9395689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9395749Z return func(*args, **kwargs) 2025-09-07T08:08:51.9395988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 294, in forward 2025-09-07T08:08:51.9396171Z query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2) 2025-09-07T08:08:51.9396381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9396495Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9396498Z 2025-09-07T08:08:51.9396572Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9396640Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9396708Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9396775Z cudagraph partition due to non gpu ops 2025-09-07T08:08:51.9396867Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9397052Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9397110Z return mod(**inputs) 2025-09-07T08:08:51.9397355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9397427Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9397671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9397731Z outputs = block( 2025-09-07T08:08:51.9397946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9398015Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9398247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9398308Z return func(*args, **kwargs) 2025-09-07T08:08:51.9398547Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9398622Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9398858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9398917Z return func(*args, **kwargs) 2025-09-07T08:08:51.9399203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9399325Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9399622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:08:51.9399740Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:08:51.9399743Z 2025-09-07T08:08:51.9399832Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9400019Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9400075Z return mod(**inputs) 2025-09-07T08:08:51.9400324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9400393Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9400696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9400758Z outputs = block( 2025-09-07T08:08:51.9400971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9401039Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9401272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9401333Z return func(*args, **kwargs) 2025-09-07T08:08:51.9401575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9401648Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9401884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9401948Z return func(*args, **kwargs) 2025-09-07T08:08:51.9402190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 336, in forward 2025-09-07T08:08:51.9402276Z attn_output, attn_weights = attention_interface( 2025-09-07T08:08:51.9402564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:08:51.9402667Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:08:51.9402670Z 2025-09-07T08:08:51.9402761Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9402946Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9403002Z return mod(**inputs) 2025-09-07T08:08:51.9403373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9403449Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9403696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9403756Z outputs = block( 2025-09-07T08:08:51.9403967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9404035Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9404268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9404327Z return func(*args, **kwargs) 2025-09-07T08:08:51.9404568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 404, in forward 2025-09-07T08:08:51.9404641Z attn_output, self_attn_weights = self.attn( 2025-09-07T08:08:51.9404876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9405034Z return func(*args, **kwargs) 2025-09-07T08:08:51.9405327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 349, in forward 2025-09-07T08:08:51.9405404Z attn_output = self.c_proj(attn_output) 2025-09-07T08:08:51.9405617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9405722Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9405725Z 2025-09-07T08:08:51.9405813Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9405997Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9406058Z return mod(**inputs) 2025-09-07T08:08:51.9406303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9406376Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9406708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9406770Z outputs = block( 2025-09-07T08:08:51.9406984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9407052Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9407293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9407352Z return func(*args, **kwargs) 2025-09-07T08:08:51.9407600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9407695Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9407936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 365, in forward 2025-09-07T08:08:51.9408010Z hidden_states = self.c_fc(hidden_states) 2025-09-07T08:08:51.9408227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9408335Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9408338Z 2025-09-07T08:08:51.9408427Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9408616Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9408674Z return mod(**inputs) 2025-09-07T08:08:51.9408920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9408993Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9409232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9409296Z outputs = block( 2025-09-07T08:08:51.9409511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9409579Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9409814Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9409875Z return func(*args, **kwargs) 2025-09-07T08:08:51.9410118Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9410207Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9410448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 366, in forward 2025-09-07T08:08:51.9410524Z hidden_states = self.act(hidden_states) 2025-09-07T08:08:51.9410728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T08:08:51.9410970Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T08:08:51.9410973Z 2025-09-07T08:08:51.9411063Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9411251Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9411308Z return mod(**inputs) 2025-09-07T08:08:51.9411555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1183, in forward 2025-09-07T08:08:51.9411628Z transformer_outputs = self.transformer( 2025-09-07T08:08:51.9411867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 917, in forward 2025-09-07T08:08:51.9411928Z outputs = block( 2025-09-07T08:08:51.9412142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:08:51.9412213Z return super().__call__(*args, **kwargs) 2025-09-07T08:08:51.9412509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:08:51.9412571Z return func(*args, **kwargs) 2025-09-07T08:08:51.9412816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 440, in forward 2025-09-07T08:08:51.9412905Z feed_forward_hidden_states = self.mlp(hidden_states) 2025-09-07T08:08:51.9413142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 367, in forward 2025-09-07T08:08:51.9413228Z hidden_states = self.c_proj(hidden_states) 2025-09-07T08:08:51.9413438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 116, in forward 2025-09-07T08:08:51.9413542Z x = torch.addmm(self.bias, x.view(-1, x.size(-1)), self.weight) 2025-09-07T08:08:51.9413547Z 2025-09-07T08:08:51.9413634Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:08:51.9413826Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:08:51.9413881Z return mod(**inputs) 2025-09-07T08:08:51.9414123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/gpt2/modeling_gpt2.py", line 1207, in forward 2025-09-07T08:08:51.9414232Z logits = self.lm_head(hidden_states[:, slice_indices, :]) 2025-09-07T08:08:51.9414235Z 2025-09-07T08:09:06.6326642Z pass 2025-09-07T08:09:06.6327068Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:09:09.2948071Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:09:09.2948984Z import pynvml # type: ignore[import] 2025-09-07T08:09:11.3088480Z 2025-09-07T08:09:25.5842927Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:09:25.5843564Z loading model: 0it [00:14, ?it/s] 2025-09-07T08:09:25.6468742Z cpu eval hf_GPT2_large 2025-09-07T08:09:26.0351735Z pass_due_to_skip 2025-09-07T08:09:26.0355607Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:09:28.4600820Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:09:28.4601696Z import pynvml # type: ignore[import] 2025-09-07T08:09:30.4714070Z 2025-09-07T08:09:34.4971958Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:09:34.4972331Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:09:34.5272628Z cpu eval hf_Longformer 2025-09-07T08:09:36.2713179Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:09:37.0645809Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:09:37.8520482Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:09:38.1032164Z W0907 08:09:38.102562 54556 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] Graph break from `Tensor.item()`, consider setting: 2025-09-07T08:09:38.1032848Z W0907 08:09:38.102562 54556 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] torch._dynamo.config.capture_scalar_outputs = True 2025-09-07T08:09:38.1033369Z W0907 08:09:38.102562 54556 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] or: 2025-09-07T08:09:38.1033852Z W0907 08:09:38.102562 54556 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1 2025-09-07T08:09:38.1034863Z W0907 08:09:38.102562 54556 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] to include these operations in the captured graph. 2025-09-07T08:09:38.1035354Z W0907 08:09:38.102562 54556 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:09:38.1035805Z W0907 08:09:38.102562 54556 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] Graph break: from user code at: 2025-09-07T08:09:38.1036477Z W0907 08:09:38.102562 54556 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:09:38.1037120Z W0907 08:09:38.102562 54556 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] return mod(**inputs) 2025-09-07T08:09:38.1037863Z W0907 08:09:38.102562 54556 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1703, in forward 2025-09-07T08:09:38.1038606Z W0907 08:09:38.102562 54556 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] outputs = self.longformer( 2025-09-07T08:09:38.1039365Z W0907 08:09:38.102562 54556 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1600, in forward 2025-09-07T08:09:38.1040146Z W0907 08:09:38.102562 54556 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] encoder_outputs = self.encoder( 2025-09-07T08:09:38.1040911Z W0907 08:09:38.102562 54556 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1244, in forward 2025-09-07T08:09:38.1041709Z W0907 08:09:38.102562 54556 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] is_global_attn = is_index_global_attn.flatten().any().item() 2025-09-07T08:09:38.1042212Z W0907 08:09:38.102562 54556 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:09:38.1042577Z W0907 08:09:38.102562 54556 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:09:44.9393034Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:09:44.9393604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1596, in forward 2025-09-07T08:09:44.9394084Z embedding_output = self.embeddings( 2025-09-07T08:09:44.9394543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 410, in forward 2025-09-07T08:09:44.9395104Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx).to(input_ids.device) 2025-09-07T08:09:44.9395742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 381, in create_position_ids_from_input_ids 2025-09-07T08:09:44.9396707Z mask = input_ids.ne(padding_idx).int() 2025-09-07T08:09:44.9396845Z 2025-09-07T08:09:44.9396960Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:09:44.9397417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1592, in forward 2025-09-07T08:09:44.9397978Z extended_attention_mask: torch.Tensor = self.get_extended_attention_mask(attention_mask, input_shape)[ 2025-09-07T08:09:44.9398525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_utils.py", line 1763, in get_extended_attention_mask 2025-09-07T08:09:44.9399020Z extended_attention_mask = (1.0 - extended_attention_mask) * torch.finfo(dtype).min 2025-09-07T08:09:44.9399248Z 2025-09-07T08:09:44.9399362Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:09:44.9399841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1596, in forward 2025-09-07T08:09:44.9400406Z embedding_output = self.embeddings( 2025-09-07T08:09:44.9400840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 410, in forward 2025-09-07T08:09:44.9401382Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx).to(input_ids.device) 2025-09-07T08:09:44.9401992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 382, in create_position_ids_from_input_ids 2025-09-07T08:09:44.9402540Z incremental_indices = torch.cumsum(mask, dim=1).type_as(mask) * mask 2025-09-07T08:09:44.9402732Z 2025-09-07T08:09:44.9402830Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:09:44.9403467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1596, in forward 2025-09-07T08:09:44.9403902Z embedding_output = self.embeddings( 2025-09-07T08:09:44.9404327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 410, in forward 2025-09-07T08:09:44.9404858Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx).to(input_ids.device) 2025-09-07T08:09:44.9405446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 382, in create_position_ids_from_input_ids 2025-09-07T08:09:44.9405992Z incremental_indices = torch.cumsum(mask, dim=1).type_as(mask) * mask 2025-09-07T08:09:44.9406184Z 2025-09-07T08:10:23.5105522Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5106176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5106740Z layer_outputs = layer_module( 2025-09-07T08:10:23.5107144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5107535Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5107972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5108409Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5108838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5109273Z self_outputs = self.self( 2025-09-07T08:10:23.5109675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:10:23.5110101Z query_vectors = self.query(hidden_states) 2025-09-07T08:10:23.5110245Z 2025-09-07T08:10:23.5110349Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5111347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5111896Z layer_outputs = layer_module( 2025-09-07T08:10:23.5112245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5112617Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5113037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5113464Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5113877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5114298Z self_outputs = self.self( 2025-09-07T08:10:23.5114702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5115273Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5115787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5116382Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5116635Z 2025-09-07T08:10:23.5116740Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5117265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5117775Z layer_outputs = layer_module( 2025-09-07T08:10:23.5118128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5118495Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5118916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5119338Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5119765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5120175Z self_outputs = self.self( 2025-09-07T08:10:23.5120574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:10:23.5121001Z key_vectors = self.key(hidden_states) 2025-09-07T08:10:23.5121137Z 2025-09-07T08:10:23.5121238Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5121756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5122258Z layer_outputs = layer_module( 2025-09-07T08:10:23.5122590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5122956Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5123367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5123782Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5124236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5124643Z self_outputs = self.self( 2025-09-07T08:10:23.5125031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5125480Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5126070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5126637Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:10:23.5127132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:10:23.5127556Z hidden_states = hidden_states.view( 2025-09-07T08:10:23.5127689Z 2025-09-07T08:10:23.5127788Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5128312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5128801Z layer_outputs = layer_module( 2025-09-07T08:10:23.5129130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5129558Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5129981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5130423Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5130825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5131238Z self_outputs = self.self( 2025-09-07T08:10:23.5131636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5132109Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5132615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5133222Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5133475Z 2025-09-07T08:10:23.5133552Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5133757Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5133955Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5134154Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5134372Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5134894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5135380Z layer_outputs = layer_module( 2025-09-07T08:10:23.5135723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5136072Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5136524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5136966Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5137405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5137882Z self_outputs = self.self( 2025-09-07T08:10:23.5146896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T08:10:23.5147417Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5147980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5148569Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T08:10:23.5149196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T08:10:23.5149779Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T08:10:23.5149985Z 2025-09-07T08:10:23.5150068Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5150316Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5150855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5151351Z layer_outputs = layer_module( 2025-09-07T08:10:23.5151716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5152086Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5152516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5153023Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5153446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5153866Z self_outputs = self.self( 2025-09-07T08:10:23.5154267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:10:23.5154681Z attn_scores += diagonal_mask 2025-09-07T08:10:23.5154800Z 2025-09-07T08:10:23.5154910Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5155429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5155934Z layer_outputs = layer_module( 2025-09-07T08:10:23.5156300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5156674Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5157150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5157586Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5158006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5158427Z self_outputs = self.self( 2025-09-07T08:10:23.5158824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:10:23.5159251Z attn_probs = nn.functional.softmax( 2025-09-07T08:10:23.5159373Z 2025-09-07T08:10:23.5159475Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5159987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5160485Z layer_outputs = layer_module( 2025-09-07T08:10:23.5160825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5161187Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5161595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5162000Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5162410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5162819Z self_outputs = self.self( 2025-09-07T08:10:23.5163214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:10:23.5163683Z value_vectors = self.value(hidden_states) 2025-09-07T08:10:23.5165296Z 2025-09-07T08:10:23.5165398Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5165913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5166408Z layer_outputs = layer_module( 2025-09-07T08:10:23.5166750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5167097Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5167505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5167927Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5168331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5168750Z self_outputs = self.self( 2025-09-07T08:10:23.5169196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5169663Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5170204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5170805Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:10:23.5171227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5171582Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5171737Z 2025-09-07T08:10:23.5171834Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5172356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5172852Z layer_outputs = layer_module( 2025-09-07T08:10:23.5173195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5173543Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5173949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5174368Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5174777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5175198Z self_outputs = self.self( 2025-09-07T08:10:23.5175599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5176072Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5176602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5177165Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:10:23.5177750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:10:23.5178214Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:10:23.5178548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5178896Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5179044Z 2025-09-07T08:10:23.5179150Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5179719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5180239Z layer_outputs = layer_module( 2025-09-07T08:10:23.5180597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5180978Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5181416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5181855Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5182265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5182676Z self_outputs = self.self( 2025-09-07T08:10:23.5183075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5183599Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5184127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5184696Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5184907Z 2025-09-07T08:10:23.5185008Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5185523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5186011Z layer_outputs = layer_module( 2025-09-07T08:10:23.5186343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5186705Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5187121Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5187536Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5187942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5188347Z self_outputs = self.self( 2025-09-07T08:10:23.5188741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5189213Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5189875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5190446Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5190655Z 2025-09-07T08:10:23.5190758Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5191278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5191762Z layer_outputs = layer_module( 2025-09-07T08:10:23.5192107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5192461Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5192863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5193298Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5193711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5194248Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5194697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5195114Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5195256Z 2025-09-07T08:10:23.5195353Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5195868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5196355Z layer_outputs = layer_module( 2025-09-07T08:10:23.5196687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5197033Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5197447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5197946Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5198359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5198806Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5199240Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5199659Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5199799Z 2025-09-07T08:10:23.5199891Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5200399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5200882Z layer_outputs = layer_module( 2025-09-07T08:10:23.5201209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5201559Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5201967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5202388Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5202783Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5203386Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5203813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5204277Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5204721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T08:10:23.5205141Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5205278Z 2025-09-07T08:10:23.5205376Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5205892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5206376Z layer_outputs = layer_module( 2025-09-07T08:10:23.5206714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5207059Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5207468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5207887Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5208368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5208819Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5209231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5209684Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5210131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T08:10:23.5210582Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:10:23.5210950Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:10:23.5211294Z return self.act(input) 2025-09-07T08:10:23.5211406Z 2025-09-07T08:10:23.5211506Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5212114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5212605Z layer_outputs = layer_module( 2025-09-07T08:10:23.5212938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5213297Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5213709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5214135Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5214527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5214918Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5215329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T08:10:23.5215797Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T08:10:23.5216252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T08:10:23.5216665Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5216803Z 2025-09-07T08:10:23.5216897Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5217405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5217964Z layer_outputs = layer_module( 2025-09-07T08:10:23.5218300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5218650Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5219067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5219486Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5219895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5220305Z self_outputs = self.self( 2025-09-07T08:10:23.5220698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:10:23.5221117Z query_vectors = self.query(hidden_states) 2025-09-07T08:10:23.5221254Z 2025-09-07T08:10:23.5221346Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5221859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5222390Z layer_outputs = layer_module( 2025-09-07T08:10:23.5222753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5223108Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5223524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5223947Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5224348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5224759Z self_outputs = self.self( 2025-09-07T08:10:23.5225158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5225602Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5226169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5226769Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5227020Z 2025-09-07T08:10:23.5227117Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5227644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5228151Z layer_outputs = layer_module( 2025-09-07T08:10:23.5228489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5228839Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5229255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5229671Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5230086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5230498Z self_outputs = self.self( 2025-09-07T08:10:23.5230884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:10:23.5231305Z key_vectors = self.key(hidden_states) 2025-09-07T08:10:23.5231444Z 2025-09-07T08:10:23.5231540Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5232055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5232541Z layer_outputs = layer_module( 2025-09-07T08:10:23.5232867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5233226Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5233639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5234052Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5234459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5234858Z self_outputs = self.self( 2025-09-07T08:10:23.5235246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5235681Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5236182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5236830Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:10:23.5237361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:10:23.5237773Z hidden_states = hidden_states.view( 2025-09-07T08:10:23.5237905Z 2025-09-07T08:10:23.5238000Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5238513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5238998Z layer_outputs = layer_module( 2025-09-07T08:10:23.5239325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5239675Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5240088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5240567Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5240971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5241382Z self_outputs = self.self( 2025-09-07T08:10:23.5241767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5242206Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5242699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5243290Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5243541Z 2025-09-07T08:10:23.5243617Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5243824Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5244027Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5244219Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5244431Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5244942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5245439Z layer_outputs = layer_module( 2025-09-07T08:10:23.5245773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5246122Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5246532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5246949Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5247363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5247776Z self_outputs = self.self( 2025-09-07T08:10:23.5248157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T08:10:23.5248604Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5249106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5249649Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T08:10:23.5250180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T08:10:23.5250711Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T08:10:23.5250967Z 2025-09-07T08:10:23.5251100Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5251356Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5251874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5252367Z layer_outputs = layer_module( 2025-09-07T08:10:23.5252702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5253059Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5253472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5253884Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5254292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5254699Z self_outputs = self.self( 2025-09-07T08:10:23.5255151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:10:23.5255570Z attn_scores += diagonal_mask 2025-09-07T08:10:23.5255687Z 2025-09-07T08:10:23.5255786Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5256290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5256773Z layer_outputs = layer_module( 2025-09-07T08:10:23.5257095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5257438Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5257914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5258331Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5258730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5259128Z self_outputs = self.self( 2025-09-07T08:10:23.5259509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:10:23.5259915Z attn_probs = nn.functional.softmax( 2025-09-07T08:10:23.5260041Z 2025-09-07T08:10:23.5260135Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5260639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5261116Z layer_outputs = layer_module( 2025-09-07T08:10:23.5261440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5261798Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5262194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5262601Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5263002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5263403Z self_outputs = self.self( 2025-09-07T08:10:23.5263781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:10:23.5264205Z value_vectors = self.value(hidden_states) 2025-09-07T08:10:23.5264341Z 2025-09-07T08:10:23.5264641Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5265138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5265696Z layer_outputs = layer_module( 2025-09-07T08:10:23.5266023Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5266367Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5266768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5267176Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5267571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5267971Z self_outputs = self.self( 2025-09-07T08:10:23.5268351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5268805Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5269391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5269973Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:10:23.5270391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5270728Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5270878Z 2025-09-07T08:10:23.5270971Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5271478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5271959Z layer_outputs = layer_module( 2025-09-07T08:10:23.5272285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5272639Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5273047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5273456Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5273856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5274252Z self_outputs = self.self( 2025-09-07T08:10:23.5274636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5275072Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5275596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5276144Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:10:23.5276648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:10:23.5277114Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:10:23.5277437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5277779Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5277923Z 2025-09-07T08:10:23.5278022Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5278534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5279060Z layer_outputs = layer_module( 2025-09-07T08:10:23.5279418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5279813Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5280231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5280649Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5281065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5281476Z self_outputs = self.self( 2025-09-07T08:10:23.5281869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5282315Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5282902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5283479Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5283685Z 2025-09-07T08:10:23.5283784Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5284296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5284791Z layer_outputs = layer_module( 2025-09-07T08:10:23.5285123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5285480Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5285895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5286324Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5286745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5287161Z self_outputs = self.self( 2025-09-07T08:10:23.5287561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5288019Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5288552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5289109Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5289323Z 2025-09-07T08:10:23.5289424Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5289945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5290435Z layer_outputs = layer_module( 2025-09-07T08:10:23.5290780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5291140Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5291571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5291985Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5292398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5292847Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5293286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5293789Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5293935Z 2025-09-07T08:10:23.5294030Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5294539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5295023Z layer_outputs = layer_module( 2025-09-07T08:10:23.5295350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5295704Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5296111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5296529Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5296937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5297455Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5297940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5298358Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5298491Z 2025-09-07T08:10:23.5298591Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5299097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5299572Z layer_outputs = layer_module( 2025-09-07T08:10:23.5299899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5300251Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5300668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5301080Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5301479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5301875Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5302291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5302742Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5303320Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T08:10:23.5303743Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5303880Z 2025-09-07T08:10:23.5303971Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5304489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5304976Z layer_outputs = layer_module( 2025-09-07T08:10:23.5305300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5305656Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5306065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5306488Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5306881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5307269Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5307681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5308261Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5308702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T08:10:23.5309158Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:10:23.5309531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:10:23.5309872Z return self.act(input) 2025-09-07T08:10:23.5309986Z 2025-09-07T08:10:23.5310083Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5310597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5311076Z layer_outputs = layer_module( 2025-09-07T08:10:23.5311497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5311861Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5312275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5312693Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5313082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5313475Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5313882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T08:10:23.5314351Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T08:10:23.5314808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T08:10:23.5315228Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5315368Z 2025-09-07T08:10:23.5315462Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5315975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5316462Z layer_outputs = layer_module( 2025-09-07T08:10:23.5316794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5317140Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5317547Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5317962Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5318372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5318787Z self_outputs = self.self( 2025-09-07T08:10:23.5319175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:10:23.5319598Z query_vectors = self.query(hidden_states) 2025-09-07T08:10:23.5319733Z 2025-09-07T08:10:23.5319825Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5320328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5320807Z layer_outputs = layer_module( 2025-09-07T08:10:23.5321130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5321481Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5321933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5322385Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5322792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5323195Z self_outputs = self.self( 2025-09-07T08:10:23.5323591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5324029Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5324533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5325119Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5325371Z 2025-09-07T08:10:23.5325525Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5326058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5326552Z layer_outputs = layer_module( 2025-09-07T08:10:23.5326888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5327237Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5327646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5328062Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5328467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5328878Z self_outputs = self.self( 2025-09-07T08:10:23.5329264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:10:23.5329679Z key_vectors = self.key(hidden_states) 2025-09-07T08:10:23.5329812Z 2025-09-07T08:10:23.5329905Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5330413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5330894Z layer_outputs = layer_module( 2025-09-07T08:10:23.5331223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5331572Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5331976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5332390Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5332808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5333216Z self_outputs = self.self( 2025-09-07T08:10:23.5333600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5334037Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5334535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5335083Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:10:23.5335582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:10:23.5336043Z hidden_states = hidden_states.view( 2025-09-07T08:10:23.5336207Z 2025-09-07T08:10:23.5336310Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5336818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5337296Z layer_outputs = layer_module( 2025-09-07T08:10:23.5337686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5338050Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5338465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5338881Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5339288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5339702Z self_outputs = self.self( 2025-09-07T08:10:23.5340163Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5340615Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5341115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5341698Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5341948Z 2025-09-07T08:10:23.5342025Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5342227Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5342424Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5342608Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5342829Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5343342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5343830Z layer_outputs = layer_module( 2025-09-07T08:10:23.5344164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5344522Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5344933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5345345Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5345760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5346169Z self_outputs = self.self( 2025-09-07T08:10:23.5346550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T08:10:23.5347001Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5347503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5348047Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T08:10:23.5348574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T08:10:23.5349097Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T08:10:23.5349305Z 2025-09-07T08:10:23.5349379Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5349601Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5350117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5351085Z layer_outputs = layer_module( 2025-09-07T08:10:23.5351416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5351774Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5352188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5352624Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5353022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5353427Z self_outputs = self.self( 2025-09-07T08:10:23.5353811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:10:23.5354230Z attn_scores += diagonal_mask 2025-09-07T08:10:23.5354348Z 2025-09-07T08:10:23.5354528Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5355032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5355521Z layer_outputs = layer_module( 2025-09-07T08:10:23.5355854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5356207Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5356615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5357020Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5357422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5357832Z self_outputs = self.self( 2025-09-07T08:10:23.5358222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:10:23.5358635Z attn_probs = nn.functional.softmax( 2025-09-07T08:10:23.5358757Z 2025-09-07T08:10:23.5358848Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5359350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5359830Z layer_outputs = layer_module( 2025-09-07T08:10:23.5360166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5360517Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5360917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5361337Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5361739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5362149Z self_outputs = self.self( 2025-09-07T08:10:23.5362528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:10:23.5362947Z value_vectors = self.value(hidden_states) 2025-09-07T08:10:23.5363085Z 2025-09-07T08:10:23.5363175Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5363678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5364160Z layer_outputs = layer_module( 2025-09-07T08:10:23.5364482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5364907Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5365316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5365728Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5366131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5366544Z self_outputs = self.self( 2025-09-07T08:10:23.5366949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5367412Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5367949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5368609Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:10:23.5369037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5369387Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5369542Z 2025-09-07T08:10:23.5369643Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5370160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5370648Z layer_outputs = layer_module( 2025-09-07T08:10:23.5370989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5371354Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5371777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5372202Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5372612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5373014Z self_outputs = self.self( 2025-09-07T08:10:23.5373407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5373866Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5374394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5374939Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:10:23.5375447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:10:23.5375917Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:10:23.5376251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5376595Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5376741Z 2025-09-07T08:10:23.5376842Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5377353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5377928Z layer_outputs = layer_module( 2025-09-07T08:10:23.5378273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5378636Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5379102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5379568Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5379979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5380393Z self_outputs = self.self( 2025-09-07T08:10:23.5380786Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5381232Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5381758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5382326Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5382532Z 2025-09-07T08:10:23.5382638Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5383214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5383695Z layer_outputs = layer_module( 2025-09-07T08:10:23.5384030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5384387Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5384799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5385217Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5385618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5386028Z self_outputs = self.self( 2025-09-07T08:10:23.5386420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5386872Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5387394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5387943Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5388148Z 2025-09-07T08:10:23.5388241Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5388748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5389231Z layer_outputs = layer_module( 2025-09-07T08:10:23.5389563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5389916Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5390325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5390738Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5391141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5391590Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5392025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5392453Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5392595Z 2025-09-07T08:10:23.5392690Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5393256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5393778Z layer_outputs = layer_module( 2025-09-07T08:10:23.5394104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5394461Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5394870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5395283Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5395682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5396121Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5396556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5397048Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5397180Z 2025-09-07T08:10:23.5397280Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5397785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5398270Z layer_outputs = layer_module( 2025-09-07T08:10:23.5398602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5398958Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5399367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5399784Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5400188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5400586Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5401004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5401474Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5401924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T08:10:23.5402351Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5402491Z 2025-09-07T08:10:23.5402586Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5403224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5403716Z layer_outputs = layer_module( 2025-09-07T08:10:23.5404054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5404413Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5404824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5405245Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5405635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5406033Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5406444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5406898Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5407336Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T08:10:23.5407917Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:10:23.5408299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:10:23.5408642Z return self.act(input) 2025-09-07T08:10:23.5408752Z 2025-09-07T08:10:23.5408859Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5409372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5409854Z layer_outputs = layer_module( 2025-09-07T08:10:23.5410193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5410554Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5411076Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5411512Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5411912Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5412308Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5412720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T08:10:23.5413187Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T08:10:23.5413647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T08:10:23.5414063Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5414203Z 2025-09-07T08:10:23.5414299Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5414817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5415308Z layer_outputs = layer_module( 2025-09-07T08:10:23.5415639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5415999Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5416412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5416832Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5417246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5417705Z self_outputs = self.self( 2025-09-07T08:10:23.5418104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:10:23.5418532Z query_vectors = self.query(hidden_states) 2025-09-07T08:10:23.5418664Z 2025-09-07T08:10:23.5418767Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5419274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5419757Z layer_outputs = layer_module( 2025-09-07T08:10:23.5420098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5420452Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5420869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5421290Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5421751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5422198Z self_outputs = self.self( 2025-09-07T08:10:23.5422596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5423035Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5423537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5424135Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5424384Z 2025-09-07T08:10:23.5424484Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5425001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5425555Z layer_outputs = layer_module( 2025-09-07T08:10:23.5425898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5426251Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5426664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5427080Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5427490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5427896Z self_outputs = self.self( 2025-09-07T08:10:23.5428283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:10:23.5428702Z key_vectors = self.key(hidden_states) 2025-09-07T08:10:23.5428834Z 2025-09-07T08:10:23.5428941Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5429448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5429925Z layer_outputs = layer_module( 2025-09-07T08:10:23.5430263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5430620Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5431035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5431445Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5431845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5432255Z self_outputs = self.self( 2025-09-07T08:10:23.5432643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5433084Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5433583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5434133Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:10:23.5434627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:10:23.5435038Z hidden_states = hidden_states.view( 2025-09-07T08:10:23.5435161Z 2025-09-07T08:10:23.5435260Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5435764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5436346Z layer_outputs = layer_module( 2025-09-07T08:10:23.5436682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5437053Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5437465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5437881Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5438283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5438691Z self_outputs = self.self( 2025-09-07T08:10:23.5439079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5439519Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5440079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5440673Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5440918Z 2025-09-07T08:10:23.5440992Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5441186Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5441379Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5441564Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5441776Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5442275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5442758Z layer_outputs = layer_module( 2025-09-07T08:10:23.5443084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5443428Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5443834Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5444243Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5453681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5454140Z self_outputs = self.self( 2025-09-07T08:10:23.5454570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T08:10:23.5455057Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5455605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5456174Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T08:10:23.5456719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T08:10:23.5457248Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T08:10:23.5457460Z 2025-09-07T08:10:23.5457537Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5457861Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5458376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5458867Z layer_outputs = layer_module( 2025-09-07T08:10:23.5459207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5459715Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5460141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5460575Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5460995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5461403Z self_outputs = self.self( 2025-09-07T08:10:23.5461801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:10:23.5462206Z attn_scores += diagonal_mask 2025-09-07T08:10:23.5462320Z 2025-09-07T08:10:23.5462422Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5463036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5463529Z layer_outputs = layer_module( 2025-09-07T08:10:23.5463870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5464222Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5464642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5465052Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5465452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5465857Z self_outputs = self.self( 2025-09-07T08:10:23.5466251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:10:23.5466662Z attn_probs = nn.functional.softmax( 2025-09-07T08:10:23.5466785Z 2025-09-07T08:10:23.5466884Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5467390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5467872Z layer_outputs = layer_module( 2025-09-07T08:10:23.5468203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5468563Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5468967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5469382Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5469783Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5470200Z self_outputs = self.self( 2025-09-07T08:10:23.5470583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:10:23.5471009Z value_vectors = self.value(hidden_states) 2025-09-07T08:10:23.5471151Z 2025-09-07T08:10:23.5471243Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5471749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5472231Z layer_outputs = layer_module( 2025-09-07T08:10:23.5472553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5472908Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5473317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5473826Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5474226Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5474628Z self_outputs = self.self( 2025-09-07T08:10:23.5475009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5475464Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5475995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5476593Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:10:23.5477012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5477424Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5477583Z 2025-09-07T08:10:23.5477679Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5478192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5478684Z layer_outputs = layer_module( 2025-09-07T08:10:23.5479014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5479363Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5479771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5480182Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5480590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5480997Z self_outputs = self.self( 2025-09-07T08:10:23.5481384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5481825Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5482342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5482879Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:10:23.5483381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:10:23.5483838Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:10:23.5484156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5484493Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5484635Z 2025-09-07T08:10:23.5484728Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5485230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5485713Z layer_outputs = layer_module( 2025-09-07T08:10:23.5486043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5486395Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5486798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5487207Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5487657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5488107Z self_outputs = self.self( 2025-09-07T08:10:23.5488489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5488932Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5489448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5490002Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5490202Z 2025-09-07T08:10:23.5490295Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5490794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5491266Z layer_outputs = layer_module( 2025-09-07T08:10:23.5491650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5492003Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5492406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5492808Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5493203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5493602Z self_outputs = self.self( 2025-09-07T08:10:23.5493981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5494419Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5494937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5495486Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5495684Z 2025-09-07T08:10:23.5495773Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5496274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5496751Z layer_outputs = layer_module( 2025-09-07T08:10:23.5497077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5497420Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5497862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5498282Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5498679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5499112Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5499543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5499956Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5500089Z 2025-09-07T08:10:23.5500177Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5500671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5501144Z layer_outputs = layer_module( 2025-09-07T08:10:23.5501532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5501913Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5502310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5502711Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5503293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5503727Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5504159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5504566Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5504693Z 2025-09-07T08:10:23.5504783Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5505440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5505923Z layer_outputs = layer_module( 2025-09-07T08:10:23.5506250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5506594Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5506990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5507397Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5507790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5508176Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5508580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5509037Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5509465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T08:10:23.5509878Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5510007Z 2025-09-07T08:10:23.5510106Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5510692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5511178Z layer_outputs = layer_module( 2025-09-07T08:10:23.5511510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5511859Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5512266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5512682Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5513075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5513463Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5513867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5514314Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5514757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T08:10:23.5515212Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:10:23.5515583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:10:23.5516060Z return self.act(input) 2025-09-07T08:10:23.5516164Z 2025-09-07T08:10:23.5516262Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5516769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5517265Z layer_outputs = layer_module( 2025-09-07T08:10:23.5517599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5517946Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5518349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5518760Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5519145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5519605Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5520007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T08:10:23.5520461Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T08:10:23.5520914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T08:10:23.5521322Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5521452Z 2025-09-07T08:10:23.5521549Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5522050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5522525Z layer_outputs = layer_module( 2025-09-07T08:10:23.5522852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5523198Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5523598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5524005Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5524406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5524803Z self_outputs = self.self( 2025-09-07T08:10:23.5525191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:10:23.5525599Z query_vectors = self.query(hidden_states) 2025-09-07T08:10:23.5525724Z 2025-09-07T08:10:23.5525816Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5526359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5526831Z layer_outputs = layer_module( 2025-09-07T08:10:23.5527152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5527492Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5527886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5528290Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5528682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5529078Z self_outputs = self.self( 2025-09-07T08:10:23.5529454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5529977Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5530480Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5531066Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5531308Z 2025-09-07T08:10:23.5531402Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5531903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5532375Z layer_outputs = layer_module( 2025-09-07T08:10:23.5532699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5533046Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5533711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5534129Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5534525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5534928Z self_outputs = self.self( 2025-09-07T08:10:23.5535313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:10:23.5535721Z key_vectors = self.key(hidden_states) 2025-09-07T08:10:23.5535846Z 2025-09-07T08:10:23.5535939Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5536438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5536917Z layer_outputs = layer_module( 2025-09-07T08:10:23.5537238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5537579Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5538052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5538457Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5538853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5539250Z self_outputs = self.self( 2025-09-07T08:10:23.5539625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5540046Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5540542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5541084Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:10:23.5541568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:10:23.5541968Z hidden_states = hidden_states.view( 2025-09-07T08:10:23.5542087Z 2025-09-07T08:10:23.5542177Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5542674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5543144Z layer_outputs = layer_module( 2025-09-07T08:10:23.5543467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5543910Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5544310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5544711Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5545103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5545499Z self_outputs = self.self( 2025-09-07T08:10:23.5545875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5546297Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5546781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5547430Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5547676Z 2025-09-07T08:10:23.5547756Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5547953Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5548134Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5548315Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5548525Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5549034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5549510Z layer_outputs = layer_module( 2025-09-07T08:10:23.5549840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5550189Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5550600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5551008Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5551408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5551818Z self_outputs = self.self( 2025-09-07T08:10:23.5552209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T08:10:23.5552309Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5552656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5552790Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T08:10:23.5553118Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T08:10:23.5553258Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T08:10:23.5553261Z 2025-09-07T08:10:23.5553332Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5553429Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5553787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5553851Z layer_outputs = layer_module( 2025-09-07T08:10:23.5554068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5554145Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5554424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5554581Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5554860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5554921Z self_outputs = self.self( 2025-09-07T08:10:23.5555202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:10:23.5555265Z attn_scores += diagonal_mask 2025-09-07T08:10:23.5555268Z 2025-09-07T08:10:23.5555358Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5555711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5555773Z layer_outputs = layer_module( 2025-09-07T08:10:23.5555988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5556118Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5556404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5556471Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5556748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5556811Z self_outputs = self.self( 2025-09-07T08:10:23.5557085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:10:23.5557157Z attn_probs = nn.functional.softmax( 2025-09-07T08:10:23.5557160Z 2025-09-07T08:10:23.5557253Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5557605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5557674Z layer_outputs = layer_module( 2025-09-07T08:10:23.5557888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5557959Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5558236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5558301Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5558577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5558635Z self_outputs = self.self( 2025-09-07T08:10:23.5558912Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:10:23.5558988Z value_vectors = self.value(hidden_states) 2025-09-07T08:10:23.5558995Z 2025-09-07T08:10:23.5559086Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5559434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5559497Z layer_outputs = layer_module( 2025-09-07T08:10:23.5559708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5559774Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5560053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5560119Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5560394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5560520Z self_outputs = self.self( 2025-09-07T08:10:23.5560797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5560909Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5561268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5561435Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:10:23.5561623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5561715Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5561718Z 2025-09-07T08:10:23.5561809Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5562252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5562320Z layer_outputs = layer_module( 2025-09-07T08:10:23.5562535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5562604Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5562880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5562945Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5563223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5563283Z self_outputs = self.self( 2025-09-07T08:10:23.5563559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5563672Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5564027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5564151Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:10:23.5564469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:10:23.5564552Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:10:23.5564734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5564819Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5564823Z 2025-09-07T08:10:23.5564913Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5565270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5565333Z layer_outputs = layer_module( 2025-09-07T08:10:23.5565546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5565614Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5565892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5565958Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5566234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5566293Z self_outputs = self.self( 2025-09-07T08:10:23.5566568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5566735Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5567090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5567234Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5567238Z 2025-09-07T08:10:23.5567328Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5567679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5567740Z layer_outputs = layer_module( 2025-09-07T08:10:23.5567958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5568024Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5568368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5568435Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5568710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5568772Z self_outputs = self.self( 2025-09-07T08:10:23.5569047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5569151Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5569500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5569638Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5569643Z 2025-09-07T08:10:23.5569734Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5570083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5570148Z layer_outputs = layer_module( 2025-09-07T08:10:23.5570358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5570431Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5570709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5570776Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5571051Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5571159Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5571437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5571510Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5571513Z 2025-09-07T08:10:23.5571605Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5571953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5572015Z layer_outputs = layer_module( 2025-09-07T08:10:23.5572225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5572288Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5572565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5572700Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5572981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5573077Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5573353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5573425Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5573428Z 2025-09-07T08:10:23.5573516Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5573867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5573929Z layer_outputs = layer_module( 2025-09-07T08:10:23.5574199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5574268Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5574543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5574618Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5574880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5574949Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5575231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5575332Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5575610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T08:10:23.5575685Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5575688Z 2025-09-07T08:10:23.5575779Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5576129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5576192Z layer_outputs = layer_module( 2025-09-07T08:10:23.5576402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5576466Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5576743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5576813Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5577074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5577139Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5577420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5577511Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5577848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T08:10:23.5577953Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:10:23.5578162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:10:23.5578225Z return self.act(input) 2025-09-07T08:10:23.5578228Z 2025-09-07T08:10:23.5578317Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5578707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5578806Z layer_outputs = layer_module( 2025-09-07T08:10:23.5579019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5579089Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5579364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5579436Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5579691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5579754Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5580034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T08:10:23.5580219Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T08:10:23.5580503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T08:10:23.5580573Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5580576Z 2025-09-07T08:10:23.5580665Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5581015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5581077Z layer_outputs = layer_module( 2025-09-07T08:10:23.5581290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5581354Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5581637Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5581706Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5581981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5582048Z self_outputs = self.self( 2025-09-07T08:10:23.5582324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:10:23.5582397Z query_vectors = self.query(hidden_states) 2025-09-07T08:10:23.5582400Z 2025-09-07T08:10:23.5582487Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5582837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5582899Z layer_outputs = layer_module( 2025-09-07T08:10:23.5583112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5583184Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5583467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5583542Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5583818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5583878Z self_outputs = self.self( 2025-09-07T08:10:23.5584159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5584253Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5584606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5584856Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5584859Z 2025-09-07T08:10:23.5584958Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5585308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5585371Z layer_outputs = layer_module( 2025-09-07T08:10:23.5585590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5585657Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5585943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5586012Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5586364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5586432Z self_outputs = self.self( 2025-09-07T08:10:23.5586714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:10:23.5586791Z key_vectors = self.key(hidden_states) 2025-09-07T08:10:23.5586794Z 2025-09-07T08:10:23.5586882Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5587233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5587296Z layer_outputs = layer_module( 2025-09-07T08:10:23.5587509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5587584Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5587863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5587934Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5588210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5588273Z self_outputs = self.self( 2025-09-07T08:10:23.5588553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5588644Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5588993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5589145Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:10:23.5589430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:10:23.5589498Z hidden_states = hidden_states.view( 2025-09-07T08:10:23.5589501Z 2025-09-07T08:10:23.5589589Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5589948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5590011Z layer_outputs = layer_module( 2025-09-07T08:10:23.5590225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5590292Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5590575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5590714Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5590988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5591056Z self_outputs = self.self( 2025-09-07T08:10:23.5591331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5591425Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5591767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5591941Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5591951Z 2025-09-07T08:10:23.5592026Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5592095Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5592230Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5592301Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5592396Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5592747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5592810Z layer_outputs = layer_module( 2025-09-07T08:10:23.5593027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5593094Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5593376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5593444Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5593726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5593795Z self_outputs = self.self( 2025-09-07T08:10:23.5594069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T08:10:23.5594175Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5594513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5594655Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T08:10:23.5594981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T08:10:23.5595119Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T08:10:23.5595124Z 2025-09-07T08:10:23.5595200Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5595293Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5595646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5595709Z layer_outputs = layer_module( 2025-09-07T08:10:23.5595919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5595993Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5596268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5596339Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5596613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5596751Z self_outputs = self.self( 2025-09-07T08:10:23.5597030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:10:23.5597096Z attn_scores += diagonal_mask 2025-09-07T08:10:23.5597099Z 2025-09-07T08:10:23.5597193Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5597544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5597611Z layer_outputs = layer_module( 2025-09-07T08:10:23.5597826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5597896Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5598172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5598311Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5598596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5598660Z self_outputs = self.self( 2025-09-07T08:10:23.5598940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:10:23.5599010Z attn_probs = nn.functional.softmax( 2025-09-07T08:10:23.5599013Z 2025-09-07T08:10:23.5599105Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5599462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5599525Z layer_outputs = layer_module( 2025-09-07T08:10:23.5599740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5599815Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5600100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5600164Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5600440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5600512Z self_outputs = self.self( 2025-09-07T08:10:23.5600788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:10:23.5600864Z value_vectors = self.value(hidden_states) 2025-09-07T08:10:23.5600867Z 2025-09-07T08:10:23.5600956Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5601309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5601379Z layer_outputs = layer_module( 2025-09-07T08:10:23.5601598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5601680Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5601968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5602048Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5602328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5602394Z self_outputs = self.self( 2025-09-07T08:10:23.5602677Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5602831Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5603384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5603558Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:10:23.5603747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5603848Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5603852Z 2025-09-07T08:10:23.5603953Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5604319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5604388Z layer_outputs = layer_module( 2025-09-07T08:10:23.5604738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5604814Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5605097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5605172Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5605447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5605517Z self_outputs = self.self( 2025-09-07T08:10:23.5605797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5605909Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5606265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5606400Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:10:23.5606721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:10:23.5606804Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:10:23.5606994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5607082Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5607086Z 2025-09-07T08:10:23.5607182Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5607540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5607608Z layer_outputs = layer_module( 2025-09-07T08:10:23.5607836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5607907Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5608190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5608259Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5608537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5608604Z self_outputs = self.self( 2025-09-07T08:10:23.5608878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5608988Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5609345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5609640Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5609644Z 2025-09-07T08:10:23.5609738Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5610093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5610163Z layer_outputs = layer_module( 2025-09-07T08:10:23.5610381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5610453Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5610731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5610798Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5611177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5611241Z self_outputs = self.self( 2025-09-07T08:10:23.5611526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5611628Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5611987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5612122Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5612126Z 2025-09-07T08:10:23.5612219Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5612579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5612643Z layer_outputs = layer_module( 2025-09-07T08:10:23.5612863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5612931Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5613213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5613280Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5613559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5613666Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5613942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5614023Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5614029Z 2025-09-07T08:10:23.5614117Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5614464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5614528Z layer_outputs = layer_module( 2025-09-07T08:10:23.5614738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5614804Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5615078Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5615143Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5615417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5615585Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5615866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5615938Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5615943Z 2025-09-07T08:10:23.5616032Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5616379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5616442Z layer_outputs = layer_module( 2025-09-07T08:10:23.5616654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5616717Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5617058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5617133Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5617397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5617465Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5617795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5617900Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5618179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T08:10:23.5618252Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5618255Z 2025-09-07T08:10:23.5618344Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5618702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5618764Z layer_outputs = layer_module( 2025-09-07T08:10:23.5618975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5619044Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5619319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5619392Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5619647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5619713Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5619999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5620093Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5620370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T08:10:23.5620472Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:10:23.5620682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:10:23.5620745Z return self.act(input) 2025-09-07T08:10:23.5620749Z 2025-09-07T08:10:23.5620836Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5621188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5621249Z layer_outputs = layer_module( 2025-09-07T08:10:23.5621509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5621609Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5621889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5621963Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5622216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5622282Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5622562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T08:10:23.5622676Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T08:10:23.5622951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T08:10:23.5623088Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5623092Z 2025-09-07T08:10:23.5623184Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5623535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5623600Z layer_outputs = layer_module( 2025-09-07T08:10:23.5623818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5623890Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5624165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5624232Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5624513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5624579Z self_outputs = self.self( 2025-09-07T08:10:23.5624857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:10:23.5624928Z query_vectors = self.query(hidden_states) 2025-09-07T08:10:23.5624931Z 2025-09-07T08:10:23.5625018Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5625368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5625429Z layer_outputs = layer_module( 2025-09-07T08:10:23.5625642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5625706Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5625987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5626054Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5626328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5626390Z self_outputs = self.self( 2025-09-07T08:10:23.5626662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5626756Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5627098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5627279Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5627319Z 2025-09-07T08:10:23.5627442Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5627794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5627860Z layer_outputs = layer_module( 2025-09-07T08:10:23.5628069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5628136Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5628412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5628477Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5628753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5628817Z self_outputs = self.self( 2025-09-07T08:10:23.5629153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:10:23.5629226Z key_vectors = self.key(hidden_states) 2025-09-07T08:10:23.5629229Z 2025-09-07T08:10:23.5629320Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5629669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5629729Z layer_outputs = layer_module( 2025-09-07T08:10:23.5629941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5630006Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5630282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5630347Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5630626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5630690Z self_outputs = self.self( 2025-09-07T08:10:23.5630964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5631057Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5631397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5631542Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:10:23.5631820Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:10:23.5631888Z hidden_states = hidden_states.view( 2025-09-07T08:10:23.5631892Z 2025-09-07T08:10:23.5631986Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5632334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5632398Z layer_outputs = layer_module( 2025-09-07T08:10:23.5632611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5632681Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5632958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5633022Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5633298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5633411Z self_outputs = self.self( 2025-09-07T08:10:23.5633724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5633814Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5634154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5634332Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5634335Z 2025-09-07T08:10:23.5634408Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5634478Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5634544Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5634612Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5634701Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5635114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5635183Z layer_outputs = layer_module( 2025-09-07T08:10:23.5635393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5635461Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5635736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5635801Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5636084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5636147Z self_outputs = self.self( 2025-09-07T08:10:23.5636426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T08:10:23.5636534Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5636870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5637007Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T08:10:23.5637330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T08:10:23.5637470Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T08:10:23.5637473Z 2025-09-07T08:10:23.5637542Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5637634Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5637985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5638049Z layer_outputs = layer_module( 2025-09-07T08:10:23.5638261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5638327Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5638604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5638669Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5638944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5639003Z self_outputs = self.self( 2025-09-07T08:10:23.5639276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:10:23.5639379Z attn_scores += diagonal_mask 2025-09-07T08:10:23.5639410Z 2025-09-07T08:10:23.5639503Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5639854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5639913Z layer_outputs = layer_module( 2025-09-07T08:10:23.5640124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5640191Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5640470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5640541Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5640818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5640891Z self_outputs = self.self( 2025-09-07T08:10:23.5641226Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:10:23.5641301Z attn_probs = nn.functional.softmax( 2025-09-07T08:10:23.5641304Z 2025-09-07T08:10:23.5641400Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5641749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5641818Z layer_outputs = layer_module( 2025-09-07T08:10:23.5642030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5642097Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5642381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5642453Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5642736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5642798Z self_outputs = self.self( 2025-09-07T08:10:23.5643082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:10:23.5643159Z value_vectors = self.value(hidden_states) 2025-09-07T08:10:23.5643162Z 2025-09-07T08:10:23.5643255Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5643614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5643679Z layer_outputs = layer_module( 2025-09-07T08:10:23.5643893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5643965Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5644251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5644318Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5644594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5644661Z self_outputs = self.self( 2025-09-07T08:10:23.5644935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5645054Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5645408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5645642Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:10:23.5645840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5645931Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5645935Z 2025-09-07T08:10:23.5646031Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5646386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5646452Z layer_outputs = layer_module( 2025-09-07T08:10:23.5646665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5646733Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5647076Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5647146Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5647437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5647500Z self_outputs = self.self( 2025-09-07T08:10:23.5647779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5647888Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5648239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5648375Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:10:23.5648698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:10:23.5648786Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:10:23.5648968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5649052Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5649055Z 2025-09-07T08:10:23.5649151Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5649498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5649568Z layer_outputs = layer_module( 2025-09-07T08:10:23.5649779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5649847Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5650129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5650198Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5650482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5650544Z self_outputs = self.self( 2025-09-07T08:10:23.5650825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5650925Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5651277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5651426Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5651465Z 2025-09-07T08:10:23.5651556Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5651952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5652016Z layer_outputs = layer_module( 2025-09-07T08:10:23.5652234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5652300Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5652577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5652645Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5652922Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5652991Z self_outputs = self.self( 2025-09-07T08:10:23.5653346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5653454Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5653805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5653941Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5653945Z 2025-09-07T08:10:23.5654038Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5654393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5654457Z layer_outputs = layer_module( 2025-09-07T08:10:23.5654670Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5654741Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5655022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5655089Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5655370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5655472Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5655748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5655822Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5655825Z 2025-09-07T08:10:23.5655912Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5656266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5656333Z layer_outputs = layer_module( 2025-09-07T08:10:23.5656562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5656635Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5656916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5656985Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5657261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5657365Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5657675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5657844Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5657847Z 2025-09-07T08:10:23.5657936Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5658286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5658349Z layer_outputs = layer_module( 2025-09-07T08:10:23.5658562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5658635Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5658911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5658990Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5659317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5659389Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5659682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5659786Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5660068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T08:10:23.5660139Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5660142Z 2025-09-07T08:10:23.5660239Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5660594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5660658Z layer_outputs = layer_module( 2025-09-07T08:10:23.5660881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5660950Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5661233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5661305Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5661566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5661640Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5661921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5662023Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5662302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T08:10:23.5662410Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:10:23.5662616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:10:23.5662680Z return self.act(input) 2025-09-07T08:10:23.5662683Z 2025-09-07T08:10:23.5662774Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5663124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5663190Z layer_outputs = layer_module( 2025-09-07T08:10:23.5663399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5663466Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5663788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5663894Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5664154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5664222Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5664509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T08:10:23.5664624Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T08:10:23.5664903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T08:10:23.5664982Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5664986Z 2025-09-07T08:10:23.5665076Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5665499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5665569Z layer_outputs = layer_module( 2025-09-07T08:10:23.5665784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5665858Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5666132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5666200Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5666476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5666541Z self_outputs = self.self( 2025-09-07T08:10:23.5666816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:10:23.5666894Z query_vectors = self.query(hidden_states) 2025-09-07T08:10:23.5666897Z 2025-09-07T08:10:23.5666994Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5667341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5667410Z layer_outputs = layer_module( 2025-09-07T08:10:23.5667621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5667692Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5667971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5668038Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5668326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5668389Z self_outputs = self.self( 2025-09-07T08:10:23.5668669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5668760Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5669103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5669287Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5669290Z 2025-09-07T08:10:23.5669380Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5669733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5669863Z layer_outputs = layer_module( 2025-09-07T08:10:23.5670081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5670147Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5670423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5670492Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5670766Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5670832Z self_outputs = self.self( 2025-09-07T08:10:23.5671105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:10:23.5671175Z key_vectors = self.key(hidden_states) 2025-09-07T08:10:23.5671183Z 2025-09-07T08:10:23.5671335Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5671689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5671755Z layer_outputs = layer_module( 2025-09-07T08:10:23.5671969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5672037Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5672313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5672376Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5672653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5672718Z self_outputs = self.self( 2025-09-07T08:10:23.5673007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5673098Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5673447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5673593Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:10:23.5673872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:10:23.5673952Z hidden_states = hidden_states.view( 2025-09-07T08:10:23.5673956Z 2025-09-07T08:10:23.5674045Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5674398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5674466Z layer_outputs = layer_module( 2025-09-07T08:10:23.5674679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5674754Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5675033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5675107Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5675383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5675448Z self_outputs = self.self( 2025-09-07T08:10:23.5675722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5675844Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5676220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5676395Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5676398Z 2025-09-07T08:10:23.5676473Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5676541Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5676611Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5676686Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5676775Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5677134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5677202Z layer_outputs = layer_module( 2025-09-07T08:10:23.5677492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5677565Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5677843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5677913Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5678190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5678261Z self_outputs = self.self( 2025-09-07T08:10:23.5678537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T08:10:23.5678637Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5678982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5679119Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T08:10:23.5679448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T08:10:23.5679590Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T08:10:23.5679593Z 2025-09-07T08:10:23.5679671Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5679758Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5680107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5680173Z layer_outputs = layer_module( 2025-09-07T08:10:23.5680384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5680465Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5680749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5680826Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5681106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5681167Z self_outputs = self.self( 2025-09-07T08:10:23.5681448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:10:23.5681512Z attn_scores += diagonal_mask 2025-09-07T08:10:23.5681515Z 2025-09-07T08:10:23.5681605Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5681955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5682083Z layer_outputs = layer_module( 2025-09-07T08:10:23.5682301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5682369Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5682656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5682723Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5683009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5683072Z self_outputs = self.self( 2025-09-07T08:10:23.5683349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:10:23.5683435Z attn_probs = nn.functional.softmax( 2025-09-07T08:10:23.5683440Z 2025-09-07T08:10:23.5683590Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5683952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5684016Z layer_outputs = layer_module( 2025-09-07T08:10:23.5684229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5684300Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5684579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5684649Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5684924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5684996Z self_outputs = self.self( 2025-09-07T08:10:23.5685276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:10:23.5685353Z value_vectors = self.value(hidden_states) 2025-09-07T08:10:23.5685356Z 2025-09-07T08:10:23.5685455Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5685808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5685876Z layer_outputs = layer_module( 2025-09-07T08:10:23.5686088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5686155Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5686440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5686512Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5686802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5686864Z self_outputs = self.self( 2025-09-07T08:10:23.5687145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5687252Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5687608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5687778Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:10:23.5687965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5688123Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5688126Z 2025-09-07T08:10:23.5688216Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5688566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5688625Z layer_outputs = layer_module( 2025-09-07T08:10:23.5688835Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5688903Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5689178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5689243Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5689521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5689644Z self_outputs = self.self( 2025-09-07T08:10:23.5689922Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5690027Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5690388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5690514Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:10:23.5690837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:10:23.5690918Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:10:23.5691101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5691194Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5691197Z 2025-09-07T08:10:23.5691287Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5691639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5691700Z layer_outputs = layer_module( 2025-09-07T08:10:23.5691918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5691982Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5692261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5692334Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5692610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5692677Z self_outputs = self.self( 2025-09-07T08:10:23.5692951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5693050Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5693401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5693541Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5693544Z 2025-09-07T08:10:23.5693633Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5693983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5694124Z layer_outputs = layer_module( 2025-09-07T08:10:23.5694339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5694407Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5694685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5694750Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5695030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5695092Z self_outputs = self.self( 2025-09-07T08:10:23.5695370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5695477Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5695887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5696032Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5696035Z 2025-09-07T08:10:23.5696125Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5696483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5696543Z layer_outputs = layer_module( 2025-09-07T08:10:23.5696757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5696830Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5697107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5697187Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5697464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5697575Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5697908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5697985Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5697988Z 2025-09-07T08:10:23.5698084Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5698433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5698503Z layer_outputs = layer_module( 2025-09-07T08:10:23.5698723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5698793Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5699078Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5699148Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5699433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5699530Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5699811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5699882Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5699885Z 2025-09-07T08:10:23.5699970Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5700368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5700484Z layer_outputs = layer_module( 2025-09-07T08:10:23.5700697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5700762Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5701043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5701114Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5701373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5701441Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5701724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5701892Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5702174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T08:10:23.5702250Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5702260Z 2025-09-07T08:10:23.5702354Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5702709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5702784Z layer_outputs = layer_module( 2025-09-07T08:10:23.5703005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5703208Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5703512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5703592Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5703871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5703941Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5704238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5704342Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5704624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T08:10:23.5704738Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:10:23.5704948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:10:23.5705025Z return self.act(input) 2025-09-07T08:10:23.5705028Z 2025-09-07T08:10:23.5705127Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5705493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5705563Z layer_outputs = layer_module( 2025-09-07T08:10:23.5705785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5705861Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5706140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5706224Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5706483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5706722Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5707016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T08:10:23.5707132Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T08:10:23.5707417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T08:10:23.5707492Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5707495Z 2025-09-07T08:10:23.5707596Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5707948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5708016Z layer_outputs = layer_module( 2025-09-07T08:10:23.5708338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5708415Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5708703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5708774Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5709058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5709126Z self_outputs = self.self( 2025-09-07T08:10:23.5709405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:10:23.5709485Z query_vectors = self.query(hidden_states) 2025-09-07T08:10:23.5709488Z 2025-09-07T08:10:23.5709581Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5709944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5710007Z layer_outputs = layer_module( 2025-09-07T08:10:23.5710220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5710292Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5710569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5710641Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5710917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5710983Z self_outputs = self.self( 2025-09-07T08:10:23.5711264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5711363Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5711714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5711898Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5711901Z 2025-09-07T08:10:23.5712003Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5712351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5712414Z layer_outputs = layer_module( 2025-09-07T08:10:23.5712624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5712729Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5713047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5713114Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5713393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5713456Z self_outputs = self.self( 2025-09-07T08:10:23.5713731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:10:23.5713805Z key_vectors = self.key(hidden_states) 2025-09-07T08:10:23.5713809Z 2025-09-07T08:10:23.5713900Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5714251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5714376Z layer_outputs = layer_module( 2025-09-07T08:10:23.5714595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5714663Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5714941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5715008Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5715300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5715366Z self_outputs = self.self( 2025-09-07T08:10:23.5715643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5715734Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5716084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5716231Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:10:23.5716512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:10:23.5716576Z hidden_states = hidden_states.view( 2025-09-07T08:10:23.5716580Z 2025-09-07T08:10:23.5716671Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5717020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5717082Z layer_outputs = layer_module( 2025-09-07T08:10:23.5717296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5717370Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5717656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5717724Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5718009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5718070Z self_outputs = self.self( 2025-09-07T08:10:23.5718349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5718444Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5718784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5719010Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5719043Z 2025-09-07T08:10:23.5719121Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5719192Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5719271Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5719343Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5719444Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5719806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5719879Z layer_outputs = layer_module( 2025-09-07T08:10:23.5720091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5720159Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5720507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5720580Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5720867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5720933Z self_outputs = self.self( 2025-09-07T08:10:23.5721211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T08:10:23.5721319Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5721661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5721810Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T08:10:23.5722141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T08:10:23.5722288Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T08:10:23.5722292Z 2025-09-07T08:10:23.5722366Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5722461Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5722821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5722889Z layer_outputs = layer_module( 2025-09-07T08:10:23.5723110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5723179Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5723459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5723539Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5723817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5723887Z self_outputs = self.self( 2025-09-07T08:10:23.5724163Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:10:23.5724234Z attn_scores += diagonal_mask 2025-09-07T08:10:23.5724237Z 2025-09-07T08:10:23.5731451Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5731902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5731976Z layer_outputs = layer_module( 2025-09-07T08:10:23.5732224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5732435Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5732737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5732814Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5733097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5733166Z self_outputs = self.self( 2025-09-07T08:10:23.5733448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:10:23.5733523Z attn_probs = nn.functional.softmax( 2025-09-07T08:10:23.5733527Z 2025-09-07T08:10:23.5733632Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5734059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5734138Z layer_outputs = layer_module( 2025-09-07T08:10:23.5734368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5734447Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5734730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5734799Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5735080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5735146Z self_outputs = self.self( 2025-09-07T08:10:23.5735426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:10:23.5735508Z value_vectors = self.value(hidden_states) 2025-09-07T08:10:23.5735516Z 2025-09-07T08:10:23.5735616Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5735972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5736036Z layer_outputs = layer_module( 2025-09-07T08:10:23.5736256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5736326Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5736607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5736673Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5736947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5737016Z self_outputs = self.self( 2025-09-07T08:10:23.5737295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5737414Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5737841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5738023Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:10:23.5738213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5738306Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5738309Z 2025-09-07T08:10:23.5738411Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5738809Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5738908Z layer_outputs = layer_module( 2025-09-07T08:10:23.5739125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5739198Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5739474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5739543Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5739821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5739883Z self_outputs = self.self( 2025-09-07T08:10:23.5740166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5740337Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5744804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5744974Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:10:23.5745328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:10:23.5745419Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:10:23.5745622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5745717Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5745721Z 2025-09-07T08:10:23.5745827Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5746201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5746278Z layer_outputs = layer_module( 2025-09-07T08:10:23.5746531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5746608Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5746899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5746971Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5747255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5747325Z self_outputs = self.self( 2025-09-07T08:10:23.5747608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5747735Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5748101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5748258Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5748262Z 2025-09-07T08:10:23.5748362Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5748723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5748793Z layer_outputs = layer_module( 2025-09-07T08:10:23.5749012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5749151Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5749467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5749544Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5749824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5749888Z self_outputs = self.self( 2025-09-07T08:10:23.5750173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5750279Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5750642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5750778Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5750783Z 2025-09-07T08:10:23.5750909Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5751344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5751408Z layer_outputs = layer_module( 2025-09-07T08:10:23.5751626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5751695Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5751979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5752045Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5752321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5752435Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5752710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5752790Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5752794Z 2025-09-07T08:10:23.5752884Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5753238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5753300Z layer_outputs = layer_module( 2025-09-07T08:10:23.5753511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5753580Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5753853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5753925Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5754199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5754295Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5754570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5754644Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5754647Z 2025-09-07T08:10:23.5754738Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5755090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5755188Z layer_outputs = layer_module( 2025-09-07T08:10:23.5755402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5755502Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5755783Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5755856Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5756120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5756189Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5756472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5756575Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5756853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T08:10:23.5756981Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5756984Z 2025-09-07T08:10:23.5757077Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5757475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5757538Z layer_outputs = layer_module( 2025-09-07T08:10:23.5757757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5757825Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5758099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5758173Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5758434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5758507Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5758792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5758886Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5759163Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T08:10:23.5759266Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:10:23.5759485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:10:23.5759549Z return self.act(input) 2025-09-07T08:10:23.5759552Z 2025-09-07T08:10:23.5759643Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5759998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5760061Z layer_outputs = layer_module( 2025-09-07T08:10:23.5760274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5760341Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5760625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5760705Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5760962Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5761037Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5761319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T08:10:23.5761504Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T08:10:23.5761781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T08:10:23.5761853Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5761857Z 2025-09-07T08:10:23.5761945Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5762298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5762361Z layer_outputs = layer_module( 2025-09-07T08:10:23.5762572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5762639Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5762950Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5763020Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5763336Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5763400Z self_outputs = self.self( 2025-09-07T08:10:23.5763695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:10:23.5763767Z query_vectors = self.query(hidden_states) 2025-09-07T08:10:23.5763770Z 2025-09-07T08:10:23.5763862Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5764214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5764278Z layer_outputs = layer_module( 2025-09-07T08:10:23.5764496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5764563Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5764843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5764911Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5765188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5765248Z self_outputs = self.self( 2025-09-07T08:10:23.5765523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5765622Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5765968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5766163Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5766167Z 2025-09-07T08:10:23.5766262Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5766620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5766684Z layer_outputs = layer_module( 2025-09-07T08:10:23.5766898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5766975Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5767253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5767363Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5767686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5767750Z self_outputs = self.self( 2025-09-07T08:10:23.5768035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:10:23.5768109Z key_vectors = self.key(hidden_states) 2025-09-07T08:10:23.5768112Z 2025-09-07T08:10:23.5768209Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5768561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5768634Z layer_outputs = layer_module( 2025-09-07T08:10:23.5768849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5768918Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5769238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5769307Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5769628Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5769693Z self_outputs = self.self( 2025-09-07T08:10:23.5769970Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5770067Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5770411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5770568Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:10:23.5770848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:10:23.5770917Z hidden_states = hidden_states.view( 2025-09-07T08:10:23.5770920Z 2025-09-07T08:10:23.5771013Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5771364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5771433Z layer_outputs = layer_module( 2025-09-07T08:10:23.5771648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5771718Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5771995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5772066Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5772343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5772403Z self_outputs = self.self( 2025-09-07T08:10:23.5772682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5772769Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5773109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5773284Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5773288Z 2025-09-07T08:10:23.5773361Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5773466Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5773566Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5773637Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5773726Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5774081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5774148Z layer_outputs = layer_module( 2025-09-07T08:10:23.5774360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5774429Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5774705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5774775Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5775087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5775151Z self_outputs = self.self( 2025-09-07T08:10:23.5775429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T08:10:23.5775567Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5775916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5776055Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T08:10:23.5776380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T08:10:23.5776528Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T08:10:23.5776534Z 2025-09-07T08:10:23.5776608Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5776706Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5777061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5777127Z layer_outputs = layer_module( 2025-09-07T08:10:23.5777344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5777412Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5777747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5777815Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5778095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5778159Z self_outputs = self.self( 2025-09-07T08:10:23.5778438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:10:23.5778505Z attn_scores += diagonal_mask 2025-09-07T08:10:23.5778508Z 2025-09-07T08:10:23.5778600Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5778953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5779014Z layer_outputs = layer_module( 2025-09-07T08:10:23.5779232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5779298Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5779577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5779731Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5780010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5780075Z self_outputs = self.self( 2025-09-07T08:10:23.5780350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:10:23.5780422Z attn_probs = nn.functional.softmax( 2025-09-07T08:10:23.5780425Z 2025-09-07T08:10:23.5780516Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5780865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5780936Z layer_outputs = layer_module( 2025-09-07T08:10:23.5781150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5781259Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5781540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5781644Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5781932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5781997Z self_outputs = self.self( 2025-09-07T08:10:23.5782280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:10:23.5782358Z value_vectors = self.value(hidden_states) 2025-09-07T08:10:23.5782361Z 2025-09-07T08:10:23.5782457Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5782807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5782874Z layer_outputs = layer_module( 2025-09-07T08:10:23.5783087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5783157Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5783443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5783511Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5783787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5783855Z self_outputs = self.self( 2025-09-07T08:10:23.5784131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5784249Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5784609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5784788Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:10:23.5784983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5785076Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5785079Z 2025-09-07T08:10:23.5785176Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5785527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5785597Z layer_outputs = layer_module( 2025-09-07T08:10:23.5785851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5785953Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5786232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5786302Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5786589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5786652Z self_outputs = self.self( 2025-09-07T08:10:23.5786933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5787037Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5787394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5787570Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:10:23.5787936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:10:23.5788019Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:10:23.5788200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5788286Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5788289Z 2025-09-07T08:10:23.5788378Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5788731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5788796Z layer_outputs = layer_module( 2025-09-07T08:10:23.5789011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5789078Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5789361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5789428Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5789703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5789765Z self_outputs = self.self( 2025-09-07T08:10:23.5790039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5790141Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5790493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5790638Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5790642Z 2025-09-07T08:10:23.5790732Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5791082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5791145Z layer_outputs = layer_module( 2025-09-07T08:10:23.5791357Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5791422Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5791699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5791799Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5792111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5792171Z self_outputs = self.self( 2025-09-07T08:10:23.5792447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5792547Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5792895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5793030Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5793033Z 2025-09-07T08:10:23.5793122Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5793505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5793569Z layer_outputs = layer_module( 2025-09-07T08:10:23.5793783Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5793882Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5794162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5794232Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5794508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5794611Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5794885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5794967Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5794975Z 2025-09-07T08:10:23.5795068Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5795417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5795481Z layer_outputs = layer_module( 2025-09-07T08:10:23.5795691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5795758Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5796034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5796098Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5796377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5796478Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5796758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5796832Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5796835Z 2025-09-07T08:10:23.5796924Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5797272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5797331Z layer_outputs = layer_module( 2025-09-07T08:10:23.5797543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5797609Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5797932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5798039Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5798303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5798380Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5798663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5798770Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5799056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T08:10:23.5799129Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5799132Z 2025-09-07T08:10:23.5799222Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5799607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5799673Z layer_outputs = layer_module( 2025-09-07T08:10:23.5799919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5799995Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5800272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5800344Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5800604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5800669Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5800959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5801058Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5801339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T08:10:23.5801442Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:10:23.5801649Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:10:23.5801712Z return self.act(input) 2025-09-07T08:10:23.5801716Z 2025-09-07T08:10:23.5801804Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5802154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5802218Z layer_outputs = layer_module( 2025-09-07T08:10:23.5802442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5802509Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5802788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5802865Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5803314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5803387Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5803669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T08:10:23.5803781Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T08:10:23.5804067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T08:10:23.5804261Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5804264Z 2025-09-07T08:10:23.5804361Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5804714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5804781Z layer_outputs = layer_module( 2025-09-07T08:10:23.5804994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5805062Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5805343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5805411Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5805738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5805806Z self_outputs = self.self( 2025-09-07T08:10:23.5806083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:10:23.5806207Z query_vectors = self.query(hidden_states) 2025-09-07T08:10:23.5806211Z 2025-09-07T08:10:23.5806300Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5806655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5806715Z layer_outputs = layer_module( 2025-09-07T08:10:23.5806929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5806995Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5807274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5807352Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5807634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5807701Z self_outputs = self.self( 2025-09-07T08:10:23.5807975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5808066Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5808412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5808593Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5808598Z 2025-09-07T08:10:23.5808693Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5809041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5809110Z layer_outputs = layer_module( 2025-09-07T08:10:23.5809324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5809396Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5809676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5809743Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5810026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5810122Z self_outputs = self.self( 2025-09-07T08:10:23.5810438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:10:23.5810509Z key_vectors = self.key(hidden_states) 2025-09-07T08:10:23.5810513Z 2025-09-07T08:10:23.5810605Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5810957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5811017Z layer_outputs = layer_module( 2025-09-07T08:10:23.5811234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5811300Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5811586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5811655Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5811967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5812033Z self_outputs = self.self( 2025-09-07T08:10:23.5812371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5812463Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5812804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5812952Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:10:23.5813235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:10:23.5813304Z hidden_states = hidden_states.view( 2025-09-07T08:10:23.5813309Z 2025-09-07T08:10:23.5813406Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5813756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5813826Z layer_outputs = layer_module( 2025-09-07T08:10:23.5814040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5814106Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5814386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5814453Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5814734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5814797Z self_outputs = self.self( 2025-09-07T08:10:23.5815075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5815166Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5815507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5815685Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5815688Z 2025-09-07T08:10:23.5815760Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5815833Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5815901Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5815971Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5816066Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5816458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5816559Z layer_outputs = layer_module( 2025-09-07T08:10:23.5816771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5816837Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5817115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5817179Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5817459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5817520Z self_outputs = self.self( 2025-09-07T08:10:23.5817848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T08:10:23.5817993Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5818368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5818512Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T08:10:23.5818838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T08:10:23.5818980Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T08:10:23.5818983Z 2025-09-07T08:10:23.5819054Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5819147Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5819501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5819568Z layer_outputs = layer_module( 2025-09-07T08:10:23.5819791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5819862Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5820149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5820217Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5820494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5820558Z self_outputs = self.self( 2025-09-07T08:10:23.5820835Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:10:23.5820902Z attn_scores += diagonal_mask 2025-09-07T08:10:23.5820906Z 2025-09-07T08:10:23.5820995Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5821345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5821408Z layer_outputs = layer_module( 2025-09-07T08:10:23.5821621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5821689Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5821964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5822030Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5822305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5822433Z self_outputs = self.self( 2025-09-07T08:10:23.5822715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:10:23.5822784Z attn_probs = nn.functional.softmax( 2025-09-07T08:10:23.5822787Z 2025-09-07T08:10:23.5822878Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5823226Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5823289Z layer_outputs = layer_module( 2025-09-07T08:10:23.5823502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5823568Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5823848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5823948Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5824234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5824295Z self_outputs = self.self( 2025-09-07T08:10:23.5824621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:10:23.5824703Z value_vectors = self.value(hidden_states) 2025-09-07T08:10:23.5824707Z 2025-09-07T08:10:23.5824793Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5825146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5825208Z layer_outputs = layer_module( 2025-09-07T08:10:23.5825423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5825493Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5825770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5825841Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5826116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5826181Z self_outputs = self.self( 2025-09-07T08:10:23.5826462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5826571Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5826929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5827100Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:10:23.5827291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5827383Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5827387Z 2025-09-07T08:10:23.5827486Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5827838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5827899Z layer_outputs = layer_module( 2025-09-07T08:10:23.5828114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5828184Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5828508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5828613Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5828900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5828964Z self_outputs = self.self( 2025-09-07T08:10:23.5829245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5829355Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5829708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5829836Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:10:23.5830188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:10:23.5830275Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:10:23.5830458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5830574Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5830578Z 2025-09-07T08:10:23.5830670Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5831021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5831084Z layer_outputs = layer_module( 2025-09-07T08:10:23.5831295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5831361Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5831641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5831708Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5831986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5832046Z self_outputs = self.self( 2025-09-07T08:10:23.5832317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5832419Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5832770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5832915Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5832920Z 2025-09-07T08:10:23.5833009Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5833366Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5833426Z layer_outputs = layer_module( 2025-09-07T08:10:23.5833640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5833714Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5833990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5834059Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5834336Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5834437Z self_outputs = self.self( 2025-09-07T08:10:23.5834716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5834849Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5835208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5835341Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5835345Z 2025-09-07T08:10:23.5835435Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5835784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5835850Z layer_outputs = layer_module( 2025-09-07T08:10:23.5836059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5836160Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5836449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5836547Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5836829Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5836930Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5837214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5837292Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5837295Z 2025-09-07T08:10:23.5837383Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5837739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5837804Z layer_outputs = layer_module( 2025-09-07T08:10:23.5838018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5838084Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5838360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5838427Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5838700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5838800Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5839074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5839149Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5839154Z 2025-09-07T08:10:23.5839242Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5839590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5839653Z layer_outputs = layer_module( 2025-09-07T08:10:23.5839865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5839937Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5840213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5840288Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5840594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5840696Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5840988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5841092Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5841376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T08:10:23.5841447Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5841450Z 2025-09-07T08:10:23.5841544Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5841896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5841961Z layer_outputs = layer_module( 2025-09-07T08:10:23.5842208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5842276Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5842590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5842666Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5842926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5842996Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5843276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5843373Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5843649Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T08:10:23.5843755Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:10:23.5843967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:10:23.5844029Z return self.act(input) 2025-09-07T08:10:23.5844033Z 2025-09-07T08:10:23.5844125Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5844472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5844534Z layer_outputs = layer_module( 2025-09-07T08:10:23.5844747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5844813Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5845094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5845166Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5845426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5845494Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5845774Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T08:10:23.5845891Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T08:10:23.5846168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T08:10:23.5846244Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5846247Z 2025-09-07T08:10:23.5846334Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5846729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5846822Z layer_outputs = layer_module( 2025-09-07T08:10:23.5847036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5847105Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5847380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5847449Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5847722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5847784Z self_outputs = self.self( 2025-09-07T08:10:23.5848062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T08:10:23.5848182Z query_vectors = self.query(hidden_states) 2025-09-07T08:10:23.5848185Z 2025-09-07T08:10:23.5848276Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5848669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5848733Z layer_outputs = layer_module( 2025-09-07T08:10:23.5848945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5849010Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5849289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5849356Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5849637Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5849698Z self_outputs = self.self( 2025-09-07T08:10:23.5849972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5850068Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5850409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5850591Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5850594Z 2025-09-07T08:10:23.5850685Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5851036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5851101Z layer_outputs = layer_module( 2025-09-07T08:10:23.5851312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5851388Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5851665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5851738Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5852014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5852083Z self_outputs = self.self( 2025-09-07T08:10:23.5852359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T08:10:23.5852429Z key_vectors = self.key(hidden_states) 2025-09-07T08:10:23.5852466Z 2025-09-07T08:10:23.5852596Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5852950Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5853018Z layer_outputs = layer_module( 2025-09-07T08:10:23.5853232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5853301Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5853586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5853652Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5853932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5853996Z self_outputs = self.self( 2025-09-07T08:10:23.5854311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5854404Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5854781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 790, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5854936Z key = self._chunk(key, window_overlap, getattr(self.config, "onnx_export", False)) 2025-09-07T08:10:23.5855218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 719, in _chunk 2025-09-07T08:10:23.5855296Z hidden_states = hidden_states.view( 2025-09-07T08:10:23.5855299Z 2025-09-07T08:10:23.5855389Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5855748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5855817Z layer_outputs = layer_module( 2025-09-07T08:10:23.5856032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5856104Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5856380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5856453Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5856731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5856792Z self_outputs = self.self( 2025-09-07T08:10:23.5857071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T08:10:23.5857162Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5857516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5857762Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T08:10:23.5857766Z 2025-09-07T08:10:23.5857842Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5857911Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5857979Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5858053Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5858142Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5858497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5858598Z layer_outputs = layer_module( 2025-09-07T08:10:23.5858815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5858919Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5859199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5859271Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5859547Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5859608Z self_outputs = self.self( 2025-09-07T08:10:23.5859890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T08:10:23.5859990Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T08:10:23.5860366Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T08:10:23.5860511Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T08:10:23.5860875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T08:10:23.5861014Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T08:10:23.5861017Z 2025-09-07T08:10:23.5861088Z cudagraph partition due to non gpu ops 2025-09-07T08:10:23.5861183Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5861536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5861604Z layer_outputs = layer_module( 2025-09-07T08:10:23.5861819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5861899Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5862179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5862247Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5862531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5862591Z self_outputs = self.self( 2025-09-07T08:10:23.5862870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T08:10:23.5862936Z attn_scores += diagonal_mask 2025-09-07T08:10:23.5862939Z 2025-09-07T08:10:23.5863028Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5863382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5863447Z layer_outputs = layer_module( 2025-09-07T08:10:23.5863664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5863733Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5864016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5864083Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5864360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5864426Z self_outputs = self.self( 2025-09-07T08:10:23.5864698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T08:10:23.5864810Z attn_probs = nn.functional.softmax( 2025-09-07T08:10:23.5864846Z 2025-09-07T08:10:23.5864937Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5865287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5865355Z layer_outputs = layer_module( 2025-09-07T08:10:23.5865566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5865641Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5865918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5865988Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5866263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5866327Z self_outputs = self.self( 2025-09-07T08:10:23.5866641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T08:10:23.5866719Z value_vectors = self.value(hidden_states) 2025-09-07T08:10:23.5866750Z 2025-09-07T08:10:23.5866848Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5867198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5867263Z layer_outputs = layer_module( 2025-09-07T08:10:23.5867476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5867542Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5867825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5867895Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5868180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5868242Z self_outputs = self.self( 2025-09-07T08:10:23.5868516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5868632Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5868986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5869156Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T08:10:23.5869345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5869443Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5869446Z 2025-09-07T08:10:23.5869539Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5869896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5869968Z layer_outputs = layer_module( 2025-09-07T08:10:23.5870190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5870274Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5870560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5870632Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5870916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5871061Z self_outputs = self.self( 2025-09-07T08:10:23.5871352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5871469Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5871830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5871961Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T08:10:23.5872282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T08:10:23.5872368Z chunked_hidden_states = nn.functional.pad( 2025-09-07T08:10:23.5872556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T08:10:23.5872691Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T08:10:23.5872694Z 2025-09-07T08:10:23.5872793Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5873182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5873250Z layer_outputs = layer_module( 2025-09-07T08:10:23.5873471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5873548Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5873831Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5873904Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5874185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5874251Z self_outputs = self.self( 2025-09-07T08:10:23.5874537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5874644Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5875000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5875147Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5875150Z 2025-09-07T08:10:23.5875247Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5875598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5875662Z layer_outputs = layer_module( 2025-09-07T08:10:23.5875884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5875954Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5876238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5876304Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5876587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:10:23.5876652Z self_outputs = self.self( 2025-09-07T08:10:23.5876929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T08:10:23.5877032Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T08:10:23.5877421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T08:10:23.5877596Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T08:10:23.5877600Z 2025-09-07T08:10:23.5877692Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5878042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5878110Z layer_outputs = layer_module( 2025-09-07T08:10:23.5878324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5878396Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5878675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5878784Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5879067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5879201Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5879486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5879563Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5879566Z 2025-09-07T08:10:23.5879664Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5880011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5880079Z layer_outputs = layer_module( 2025-09-07T08:10:23.5880295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5880364Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5880646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:10:23.5880715Z self_attn_outputs = self.attention( 2025-09-07T08:10:23.5880996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T08:10:23.5881092Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:10:23.5881369Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T08:10:23.5881447Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5881450Z 2025-09-07T08:10:23.5881543Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5881903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5881966Z layer_outputs = layer_module( 2025-09-07T08:10:23.5882185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5882255Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5882532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5882609Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5882871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5882948Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5883235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5883403Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5883685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T08:10:23.5883760Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5883763Z 2025-09-07T08:10:23.5883861Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5884213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5884280Z layer_outputs = layer_module( 2025-09-07T08:10:23.5884492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5884561Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5884880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5884954Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5885254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5885324Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5885611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T08:10:23.5885710Z intermediate_output = self.intermediate(attn_output) 2025-09-07T08:10:23.5885992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T08:10:23.5886102Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:10:23.5886311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:10:23.5886384Z return self.act(input) 2025-09-07T08:10:23.5886387Z 2025-09-07T08:10:23.5886477Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:10:23.5886828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T08:10:23.5886899Z layer_outputs = layer_module( 2025-09-07T08:10:23.5887113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:10:23.5887187Z return super().__call__(*args, **kwargs) 2025-09-07T08:10:23.5887466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T08:10:23.5887548Z layer_output = apply_chunking_to_forward( 2025-09-07T08:10:23.5887811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:10:23.5887882Z return forward_fn(*input_tensors) 2025-09-07T08:10:23.5888169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T08:10:23.5888283Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T08:10:23.5888566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T08:10:23.5888639Z hidden_states = self.dense(hidden_states) 2025-09-07T08:10:23.5888642Z 2025-09-07T08:11:27.7024171Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:27.7024831Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1716, in torch_dynamo_resume_in_forward_at_1703 2025-09-07T08:11:27.7025865Z prediction_scores = self.lm_head(sequence_output) 2025-09-07T08:11:27.7026476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1333, in forward 2025-09-07T08:11:27.7026922Z x = self.dense(features) 2025-09-07T08:11:27.7027037Z 2025-09-07T08:11:27.7027152Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:27.7027698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1716, in torch_dynamo_resume_in_forward_at_1703 2025-09-07T08:11:27.7028222Z prediction_scores = self.lm_head(sequence_output) 2025-09-07T08:11:27.7028679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1338, in forward 2025-09-07T08:11:27.7029091Z x = self.decoder(x) 2025-09-07T08:11:27.7029192Z 2025-09-07T08:11:30.2132580Z pass 2025-09-07T08:11:30.2132989Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:11:34.5957358Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:11:34.5958242Z import pynvml # type: ignore[import] 2025-09-07T08:11:36.6133919Z 2025-09-07T08:11:38.0337264Z loading model: 0it [00:00, ?it/s]Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T08:11:38.0338752Z WARNING:transformers.models.reformer.modeling_reformer:Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T08:11:38.8074858Z 2025-09-07T08:11:38.8075350Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:11:38.8126092Z cpu eval hf_Reformer 2025-09-07T08:11:39.3268720Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:11:40.8352800Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:11:42.3825129Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:11:48.2491992Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:48.2492555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2095, in forward 2025-09-07T08:11:48.2493088Z least_common_mult_chunk_length = _get_least_common_mult_chunk_len(self.config) 2025-09-07T08:11:48.2493697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 194, in _get_least_common_mult_chunk_len 2025-09-07T08:11:48.2494279Z return np.lcm(config.lsh_attn_chunk_length, config.local_attn_chunk_length) 2025-09-07T08:11:48.2494494Z 2025-09-07T08:11:48.2494611Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:48.2495064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2099, in forward 2025-09-07T08:11:48.2495518Z input_shape[-1] % least_common_mult_chunk_length != 0 2025-09-07T08:11:48.2495680Z 2025-09-07T08:11:48.6720067Z W0907 08:11:48.671357 56129 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] Graph break from `Tensor.item()`, consider setting: 2025-09-07T08:11:48.6720753Z W0907 08:11:48.671357 56129 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] torch._dynamo.config.capture_scalar_outputs = True 2025-09-07T08:11:48.6721604Z W0907 08:11:48.671357 56129 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] or: 2025-09-07T08:11:48.6722203Z W0907 08:11:48.671357 56129 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1 2025-09-07T08:11:48.6722783Z W0907 08:11:48.671357 56129 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] to include these operations in the captured graph. 2025-09-07T08:11:48.6723259Z W0907 08:11:48.671357 56129 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T08:11:48.6723746Z W0907 08:11:48.671357 56129 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] Graph break: from user code at: 2025-09-07T08:11:48.6724613Z W0907 08:11:48.671357 56129 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2132, in torch_dynamo_resume_in_forward_at_2104 2025-09-07T08:11:48.6725512Z W0907 08:11:48.671357 56129 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] embedding_output = self.embeddings( 2025-09-07T08:11:48.6726373Z W0907 08:11:48.671357 56129 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 369, in forward 2025-09-07T08:11:48.6727241Z W0907 08:11:48.671357 56129 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] position_embeddings = self.position_embeddings(position_ids) 2025-09-07T08:11:48.6728039Z W0907 08:11:48.671357 56129 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 293, in forward 2025-09-07T08:11:48.6728791Z W0907 08:11:48.671357 56129 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] max_position_id = position_ids.max().item() 2025-09-07T08:11:48.6729258Z W0907 08:11:48.671357 56129 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T08:11:48.6729637Z W0907 08:11:48.671357 56129 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T08:11:48.8052588Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:48.8053159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 351, in forward 2025-09-07T08:11:48.8053599Z position_ids = torch.arange( 2025-09-07T08:11:48.8053730Z 2025-09-07T08:11:48.8053837Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:48.8054288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 357, in forward 2025-09-07T08:11:48.8054722Z inputs_embeds = self.word_embeddings(input_ids) 2025-09-07T08:11:48.8054873Z 2025-09-07T08:11:49.5570884Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:49.5571441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 293, in forward 2025-09-07T08:11:49.5571946Z max_position_id = position_ids.max().item() 2025-09-07T08:11:49.5572091Z 2025-09-07T08:11:51.2869060Z cudagraph partition due to non gpu ops 2025-09-07T08:11:51.2869347Z cudagraph partition due to non gpu ops 2025-09-07T08:11:51.2869629Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:51.2870200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 297, in torch_dynamo_resume_in_forward_at_293 2025-09-07T08:11:51.2870712Z position_encodings = torch.cat( 2025-09-07T08:11:51.2870849Z 2025-09-07T08:11:51.2870965Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:51.2871484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 303, in torch_dynamo_resume_in_forward_at_293 2025-09-07T08:11:51.2871976Z position_encodings = torch.cat( 2025-09-07T08:11:51.2872462Z 2025-09-07T08:11:52.2203406Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:52.2204403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 370, in torch_dynamo_resume_in_forward_at_369 2025-09-07T08:11:52.2204960Z embeddings = embeddings + position_embeddings 2025-09-07T08:11:52.2205126Z 2025-09-07T08:11:55.3797591Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3798227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3798763Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3799227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3799650Z layer_outputs = layer( 2025-09-07T08:11:55.3800394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3800848Z attn_outputs = self.attention( 2025-09-07T08:11:55.3801286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3801826Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3802305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:11:55.3802790Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:11:55.3803556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 749, in _hash_vectors 2025-09-07T08:11:55.3804108Z random_rotations = torch.randn(rotations_shape, device=vectors.device, dtype=vectors.dtype) 2025-09-07T08:11:55.3804362Z 2025-09-07T08:11:55.3804472Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3805003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3805510Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3805938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3806348Z layer_outputs = layer( 2025-09-07T08:11:55.3806731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3807147Z attn_outputs = self.attention( 2025-09-07T08:11:55.3807538Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3807970Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3808431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:11:55.3808973Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:11:55.3809593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 807, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:11:55.3810150Z undo_sorted_bucket_idx.scatter_(-1, sorted_bucket_idx, indices) 2025-09-07T08:11:55.3810328Z 2025-09-07T08:11:55.3810445Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3810945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3811438Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3811940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3812436Z layer_outputs = layer( 2025-09-07T08:11:55.3812821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3813242Z attn_outputs = self.attention( 2025-09-07T08:11:55.3813629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3814053Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3814471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:11:55.3814990Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:11:55.3815654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 800, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:11:55.3816216Z torch.arange(sorted_bucket_idx.shape[-1], device=buckets.device) 2025-09-07T08:11:55.3816412Z 2025-09-07T08:11:55.3816511Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3817068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3817567Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3818044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3818440Z layer_outputs = layer( 2025-09-07T08:11:55.3818817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3819229Z attn_outputs = self.attention( 2025-09-07T08:11:55.3819626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3820043Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3820459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:11:55.3820941Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:11:55.3821435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 749, in _hash_vectors 2025-09-07T08:11:55.3821975Z random_rotations = torch.randn(rotations_shape, device=vectors.device, dtype=vectors.dtype) 2025-09-07T08:11:55.3822219Z 2025-09-07T08:11:55.3822321Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3822819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3823320Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3823732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3824134Z layer_outputs = layer( 2025-09-07T08:11:55.3824509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3824909Z attn_outputs = self.attention( 2025-09-07T08:11:55.3825307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3825738Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3826179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:11:55.3826763Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:11:55.3827406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 807, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:11:55.3827961Z undo_sorted_bucket_idx.scatter_(-1, sorted_bucket_idx, indices) 2025-09-07T08:11:55.3828149Z 2025-09-07T08:11:55.3828249Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3828761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3829259Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3829682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3830082Z layer_outputs = layer( 2025-09-07T08:11:55.3830508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3830919Z attn_outputs = self.attention( 2025-09-07T08:11:55.3831355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3831790Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3832220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:11:55.3832777Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:11:55.3833414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 800, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:11:55.3833974Z torch.arange(sorted_bucket_idx.shape[-1], device=buckets.device) 2025-09-07T08:11:55.3834172Z 2025-09-07T08:11:55.3834275Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3834784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3835286Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3835701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3836101Z layer_outputs = layer( 2025-09-07T08:11:55.3836476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3836881Z attn_outputs = self.attention( 2025-09-07T08:11:55.3837265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3837686Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3838101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:11:55.3838570Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:11:55.3839064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 749, in _hash_vectors 2025-09-07T08:11:55.3839607Z random_rotations = torch.randn(rotations_shape, device=vectors.device, dtype=vectors.dtype) 2025-09-07T08:11:55.3839855Z 2025-09-07T08:11:55.3839963Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3840480Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3840967Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3841430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3841871Z layer_outputs = layer( 2025-09-07T08:11:55.3842257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3842658Z attn_outputs = self.attention( 2025-09-07T08:11:55.3843048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3843468Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3843887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:11:55.3844409Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:11:55.3845061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 807, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:11:55.3845613Z undo_sorted_bucket_idx.scatter_(-1, sorted_bucket_idx, indices) 2025-09-07T08:11:55.3845801Z 2025-09-07T08:11:55.3845933Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3846442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3846937Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3847360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3847754Z layer_outputs = layer( 2025-09-07T08:11:55.3857974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3858471Z attn_outputs = self.attention( 2025-09-07T08:11:55.3858909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3859357Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3859792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:11:55.3860325Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:11:55.3860952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 800, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:11:55.3861506Z torch.arange(sorted_bucket_idx.shape[-1], device=buckets.device) 2025-09-07T08:11:55.3861702Z 2025-09-07T08:11:55.3861811Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3862328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3862835Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3863277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3863678Z layer_outputs = layer( 2025-09-07T08:11:55.3864060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3864468Z attn_outputs = self.attention( 2025-09-07T08:11:55.3864869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3865296Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3865704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1198, in forward 2025-09-07T08:11:55.3866242Z query_vectors = self.query(hidden_states) 2025-09-07T08:11:55.3866384Z 2025-09-07T08:11:55.3866483Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3866990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3867501Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3867915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3868314Z layer_outputs = layer( 2025-09-07T08:11:55.3868690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3869094Z attn_outputs = self.attention( 2025-09-07T08:11:55.3869525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3869942Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3870391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1199, in forward 2025-09-07T08:11:55.3870805Z key_vectors = self.key(hidden_states) 2025-09-07T08:11:55.3870936Z 2025-09-07T08:11:55.3871022Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.3871221Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.3871445Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3871945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3872433Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3872841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3873236Z layer_outputs = layer( 2025-09-07T08:11:55.3873612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3874012Z attn_outputs = self.attention( 2025-09-07T08:11:55.3874403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3874819Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3875221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:11:55.3875734Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.3876267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:11:55.3876771Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:11:55.3876977Z 2025-09-07T08:11:55.3877061Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.3877267Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.3877491Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3877992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3878486Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3878889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3879280Z layer_outputs = layer( 2025-09-07T08:11:55.3879664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3881544Z attn_outputs = self.attention( 2025-09-07T08:11:55.3881944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3882377Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3882789Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:11:55.3883298Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.3883837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:11:55.3884260Z return torch.cat(slices, dim=3) 2025-09-07T08:11:55.3884384Z 2025-09-07T08:11:55.3884481Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3885023Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3885520Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3885968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3886366Z layer_outputs = layer( 2025-09-07T08:11:55.3886733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3887135Z attn_outputs = self.attention( 2025-09-07T08:11:55.3887524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3887938Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3888347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:11:55.3888828Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:11:55.3889046Z 2025-09-07T08:11:55.3889141Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3889652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3890149Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3890557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3890954Z layer_outputs = layer( 2025-09-07T08:11:55.3891332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3891739Z attn_outputs = self.attention( 2025-09-07T08:11:55.3892129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3892539Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3892947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:11:55.3893420Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:11:55.3893619Z 2025-09-07T08:11:55.3893721Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3894216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3894693Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3895114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3895588Z layer_outputs = layer( 2025-09-07T08:11:55.3896008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3896411Z attn_outputs = self.attention( 2025-09-07T08:11:55.3896797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3897211Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3897669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1200, in forward 2025-09-07T08:11:55.3898089Z value_vectors = self.value(hidden_states) 2025-09-07T08:11:55.3898223Z 2025-09-07T08:11:55.3898305Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.3898500Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.3898721Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3899275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3899784Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3900242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3900653Z layer_outputs = layer( 2025-09-07T08:11:55.3901039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3901442Z attn_outputs = self.attention( 2025-09-07T08:11:55.3901841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3902260Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3902674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1266, in forward 2025-09-07T08:11:55.3903412Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.3903965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:11:55.3904466Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:11:55.3904673Z 2025-09-07T08:11:55.3904755Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.3904960Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.3905185Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3905687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3906181Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3906597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3906996Z layer_outputs = layer( 2025-09-07T08:11:55.3907370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3907771Z attn_outputs = self.attention( 2025-09-07T08:11:55.3908161Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3908573Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3908992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:11:55.3909435Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:11:55.3909604Z 2025-09-07T08:11:55.3909707Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3910286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3910823Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3911234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3911633Z layer_outputs = layer( 2025-09-07T08:11:55.3912005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3912405Z attn_outputs = self.attention( 2025-09-07T08:11:55.3912804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3913227Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3913646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:11:55.3914156Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:11:55.3914329Z 2025-09-07T08:11:55.3914429Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3914999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3915495Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3915911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3916313Z layer_outputs = layer( 2025-09-07T08:11:55.3916680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3917080Z attn_outputs = self.attention( 2025-09-07T08:11:55.3917475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3917917Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3918326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1324, in forward 2025-09-07T08:11:55.3918858Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:11:55.3919441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:11:55.3919959Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:11:55.3920152Z 2025-09-07T08:11:55.3920255Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3920756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3921240Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3921652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3922053Z layer_outputs = layer( 2025-09-07T08:11:55.3922427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3922830Z attn_outputs = self.attention( 2025-09-07T08:11:55.3923219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T08:11:55.3923681Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T08:11:55.3924152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T08:11:55.3924647Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.3924784Z 2025-09-07T08:11:55.3925027Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3925533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3926024Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3926438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3926837Z layer_outputs = layer( 2025-09-07T08:11:55.3927210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.3927666Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.3928107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.3928555Z return apply_chunking_to_forward( 2025-09-07T08:11:55.3928960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.3929391Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.3929803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:11:55.3930242Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.3930652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T08:11:55.3931062Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.3931192Z 2025-09-07T08:11:55.3931299Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3931797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3932288Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3932704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3933103Z layer_outputs = layer( 2025-09-07T08:11:55.3933477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.3933920Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.3934367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.3934777Z return apply_chunking_to_forward( 2025-09-07T08:11:55.3935175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.3935573Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.3935981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:11:55.3936409Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.3936817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T08:11:55.3937230Z hidden_states = self.act_fn(hidden_states) 2025-09-07T08:11:55.3937366Z 2025-09-07T08:11:55.3937467Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3938026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3938522Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3938989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3939443Z layer_outputs = layer( 2025-09-07T08:11:55.3939825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.3940272Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.3940722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.3941133Z return apply_chunking_to_forward( 2025-09-07T08:11:55.3941522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.3941910Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.3942325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T08:11:55.3942790Z return self.output(hidden_states) 2025-09-07T08:11:55.3943187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T08:11:55.3943634Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.3943769Z 2025-09-07T08:11:55.3943868Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3944369Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3944863Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3945281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3945679Z layer_outputs = layer( 2025-09-07T08:11:55.3946054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3946463Z attn_outputs = self.attention( 2025-09-07T08:11:55.3946857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3947280Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3947700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 540, in forward 2025-09-07T08:11:55.3948119Z query_key_vectors = self.query_key(hidden_states) 2025-09-07T08:11:55.3948271Z 2025-09-07T08:11:55.3948365Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3948862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3949353Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3949765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3950157Z layer_outputs = layer( 2025-09-07T08:11:55.3950530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3950931Z attn_outputs = self.attention( 2025-09-07T08:11:55.3951320Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3951730Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3952145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:11:55.3952622Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:11:55.3953121Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 749, in _hash_vectors 2025-09-07T08:11:55.3953734Z random_rotations = torch.randn(rotations_shape, device=vectors.device, dtype=vectors.dtype) 2025-09-07T08:11:55.3953981Z 2025-09-07T08:11:55.3954090Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3954587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3955078Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3955492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3955890Z layer_outputs = layer( 2025-09-07T08:11:55.3956264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3956669Z attn_outputs = self.attention( 2025-09-07T08:11:55.3957093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3957523Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3957969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:11:55.3958435Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:11:55.3958926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 751, in _hash_vectors 2025-09-07T08:11:55.3959412Z rotated_vectors = torch.einsum("bmtd,mdhr->bmhtr", vectors, random_rotations) 2025-09-07T08:11:55.3959616Z 2025-09-07T08:11:55.3959714Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3960203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3960686Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3961090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3961479Z layer_outputs = layer( 2025-09-07T08:11:55.3961844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3962236Z attn_outputs = self.attention( 2025-09-07T08:11:55.3962613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3963018Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3963416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:11:55.3963876Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:11:55.3964351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T08:11:55.3964826Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T08:11:55.3965025Z 2025-09-07T08:11:55.3965120Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3965607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3966092Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3966491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3966873Z layer_outputs = layer( 2025-09-07T08:11:55.3967285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3967731Z attn_outputs = self.attention( 2025-09-07T08:11:55.3968119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3968528Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3968928Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:11:55.3969387Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:11:55.3969854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T08:11:55.3970289Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T08:11:55.3970434Z 2025-09-07T08:11:55.3970533Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3971069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3971557Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3971999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3972397Z layer_outputs = layer( 2025-09-07T08:11:55.3972765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3973158Z attn_outputs = self.attention( 2025-09-07T08:11:55.3973543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3973958Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3974364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:11:55.3974876Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:11:55.3975484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:11:55.3976005Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:11:55.3976447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T08:11:55.3976938Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T08:11:55.3977142Z 2025-09-07T08:11:55.3977222Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.3977446Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3978009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3978494Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3978910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3979301Z layer_outputs = layer( 2025-09-07T08:11:55.3979676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3980071Z attn_outputs = self.attention( 2025-09-07T08:11:55.3980461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3980874Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3981287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:11:55.3981900Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:11:55.3982511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:11:55.3983028Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:11:55.3983468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T08:11:55.3983907Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T08:11:55.3984053Z 2025-09-07T08:11:55.3984125Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.3984320Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.3984532Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3985069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3985555Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3986006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3986401Z layer_outputs = layer( 2025-09-07T08:11:55.3986788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3987204Z attn_outputs = self.attention( 2025-09-07T08:11:55.3987589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3987996Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3988424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.3988846Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.3989268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:11:55.3989808Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.3990358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:11:55.3990845Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:11:55.3991045Z 2025-09-07T08:11:55.3991119Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.3991314Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.3991526Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3992025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3992509Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3992925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.3993318Z layer_outputs = layer( 2025-09-07T08:11:55.3993682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.3994092Z attn_outputs = self.attention( 2025-09-07T08:11:55.3994475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.3994897Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.3995300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.3995872Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.3996316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:11:55.3996877Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.3997417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:11:55.3997852Z return torch.cat(slices, dim=3) 2025-09-07T08:11:55.3997984Z 2025-09-07T08:11:55.3998054Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.3998304Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.3998805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.3999297Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.3999819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4000217Z layer_outputs = layer( 2025-09-07T08:11:55.4000630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4001040Z attn_outputs = self.attention( 2025-09-07T08:11:55.4001426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4001845Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4002254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 541, in forward 2025-09-07T08:11:55.4002669Z value_vectors = self.value(hidden_states) 2025-09-07T08:11:55.4002808Z 2025-09-07T08:11:55.4002913Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4003564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4004063Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4004473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4004867Z layer_outputs = layer( 2025-09-07T08:11:55.4005240Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4005637Z attn_outputs = self.attention( 2025-09-07T08:11:55.4006030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4006457Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4006869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4007296Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4007712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:11:55.4008221Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4008747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:11:55.4009237Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:11:55.4009440Z 2025-09-07T08:11:55.4009522Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4009740Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4010325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4010877Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4011303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4011707Z layer_outputs = layer( 2025-09-07T08:11:55.4012081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4012488Z attn_outputs = self.attention( 2025-09-07T08:11:55.4012883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4013302Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4013759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4014198Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4014673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:11:55.4015189Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4015720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:11:55.4016136Z return torch.cat(slices, dim=3) 2025-09-07T08:11:55.4016266Z 2025-09-07T08:11:55.4016365Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4016869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4017359Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4017859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4018253Z layer_outputs = layer( 2025-09-07T08:11:55.4018630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4019031Z attn_outputs = self.attention( 2025-09-07T08:11:55.4019423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4019848Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4020255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T08:11:55.4020795Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T08:11:55.4021379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T08:11:55.4021846Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T08:11:55.4021991Z 2025-09-07T08:11:55.4022094Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4022587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4023075Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4023484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4023880Z layer_outputs = layer( 2025-09-07T08:11:55.4024251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4024742Z attn_outputs = self.attention( 2025-09-07T08:11:55.4025131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4025548Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4025961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4026390Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4026806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T08:11:55.4027281Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:11:55.4027492Z 2025-09-07T08:11:55.4027587Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4028130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4028627Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4029071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4029474Z layer_outputs = layer( 2025-09-07T08:11:55.4029845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4030244Z attn_outputs = self.attention( 2025-09-07T08:11:55.4030623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4031043Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4031450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4031883Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4032300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T08:11:55.4032816Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4033356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:11:55.4033849Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:11:55.4034066Z 2025-09-07T08:11:55.4034142Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4034371Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4034864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4035359Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4035775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4036175Z layer_outputs = layer( 2025-09-07T08:11:55.4036552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4036951Z attn_outputs = self.attention( 2025-09-07T08:11:55.4037340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4037766Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4038183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4038613Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4039127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:11:55.4039569Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:11:55.4039743Z 2025-09-07T08:11:55.4039841Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4040339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4040826Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4041231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4041628Z layer_outputs = layer( 2025-09-07T08:11:55.4042002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4042406Z attn_outputs = self.attention( 2025-09-07T08:11:55.4042833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4043249Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4043700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4044130Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4044555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:11:55.4044986Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:11:55.4045155Z 2025-09-07T08:11:55.4045231Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4045458Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4045955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4046444Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4046849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4047244Z layer_outputs = layer( 2025-09-07T08:11:55.4047611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4048015Z attn_outputs = self.attention( 2025-09-07T08:11:55.4048404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4048812Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4049221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T08:11:55.4049762Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:11:55.4050347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:11:55.4050864Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:11:55.4051058Z 2025-09-07T08:11:55.4051154Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4051653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4052142Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4052545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4053009Z layer_outputs = layer( 2025-09-07T08:11:55.4053375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4053776Z attn_outputs = self.attention( 2025-09-07T08:11:55.4054167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T08:11:55.4054636Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T08:11:55.4055103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T08:11:55.4055510Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4055652Z 2025-09-07T08:11:55.4055747Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4056239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4056763Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4057167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4057590Z layer_outputs = layer( 2025-09-07T08:11:55.4058002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.4058451Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.4058898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.4059301Z return apply_chunking_to_forward( 2025-09-07T08:11:55.4059684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.4060077Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.4060488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:11:55.4060916Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4061316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T08:11:55.4061726Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4061866Z 2025-09-07T08:11:55.4061965Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4062465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4062955Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4063355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4063760Z layer_outputs = layer( 2025-09-07T08:11:55.4064133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.4064582Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.4065030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.4065427Z return apply_chunking_to_forward( 2025-09-07T08:11:55.4065811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.4066205Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.4066619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:11:55.4067094Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4067523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T08:11:55.4067936Z hidden_states = self.act_fn(hidden_states) 2025-09-07T08:11:55.4068075Z 2025-09-07T08:11:55.4068169Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4068664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4069149Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4069548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4069945Z layer_outputs = layer( 2025-09-07T08:11:55.4070315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.4070796Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.4071239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.4071695Z return apply_chunking_to_forward( 2025-09-07T08:11:55.4072085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.4072479Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.4072890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T08:11:55.4073307Z return self.output(hidden_states) 2025-09-07T08:11:55.4073700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T08:11:55.4074112Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4074245Z 2025-09-07T08:11:55.4074345Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4074840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4075321Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4075743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4076135Z layer_outputs = layer( 2025-09-07T08:11:55.4076514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4076919Z attn_outputs = self.attention( 2025-09-07T08:11:55.4077301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4077722Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4078137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1198, in forward 2025-09-07T08:11:55.4078547Z query_vectors = self.query(hidden_states) 2025-09-07T08:11:55.4078676Z 2025-09-07T08:11:55.4078778Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4079264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4079748Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4080152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4080549Z layer_outputs = layer( 2025-09-07T08:11:55.4080920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4081411Z attn_outputs = self.attention( 2025-09-07T08:11:55.4081823Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4082238Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4082659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1199, in forward 2025-09-07T08:11:55.4083062Z key_vectors = self.key(hidden_states) 2025-09-07T08:11:55.4083201Z 2025-09-07T08:11:55.4083277Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4083480Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4083700Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4084197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4084675Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4085138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4085541Z layer_outputs = layer( 2025-09-07T08:11:55.4085955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4086362Z attn_outputs = self.attention( 2025-09-07T08:11:55.4086747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4087168Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4087578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:11:55.4088097Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4088628Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:11:55.4089133Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:11:55.4089347Z 2025-09-07T08:11:55.4089420Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4089625Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4089851Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4090345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4090836Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4091246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4091651Z layer_outputs = layer( 2025-09-07T08:11:55.4092022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4092415Z attn_outputs = self.attention( 2025-09-07T08:11:55.4092797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4093210Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4093621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:11:55.4094125Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4094640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:11:55.4095062Z return torch.cat(slices, dim=3) 2025-09-07T08:11:55.4095241Z 2025-09-07T08:11:55.4095368Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4095874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4096359Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4096762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4097162Z layer_outputs = layer( 2025-09-07T08:11:55.4097553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4098019Z attn_outputs = self.attention( 2025-09-07T08:11:55.4098414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4098831Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4099289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:11:55.4099771Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:11:55.4099975Z 2025-09-07T08:11:55.4100111Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4100608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4101086Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4101502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4101903Z layer_outputs = layer( 2025-09-07T08:11:55.4102273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4102673Z attn_outputs = self.attention( 2025-09-07T08:11:55.4103228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4103649Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4104064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:11:55.4104537Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:11:55.4104731Z 2025-09-07T08:11:55.4104824Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4105338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4105827Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4106247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4106650Z layer_outputs = layer( 2025-09-07T08:11:55.4107021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4107419Z attn_outputs = self.attention( 2025-09-07T08:11:55.4107805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4108223Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4108629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1200, in forward 2025-09-07T08:11:55.4109035Z value_vectors = self.value(hidden_states) 2025-09-07T08:11:55.4109174Z 2025-09-07T08:11:55.4109248Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4109548Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4109822Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4110319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4110807Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4111213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4111617Z layer_outputs = layer( 2025-09-07T08:11:55.4111985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4112377Z attn_outputs = self.attention( 2025-09-07T08:11:55.4112768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4113187Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4113645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1266, in forward 2025-09-07T08:11:55.4114228Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4114771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:11:55.4115265Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:11:55.4115478Z 2025-09-07T08:11:55.4115555Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4115759Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4115983Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4116482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4116976Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4117383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4117785Z layer_outputs = layer( 2025-09-07T08:11:55.4118156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4118549Z attn_outputs = self.attention( 2025-09-07T08:11:55.4118938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4119349Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4119759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:11:55.4120199Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:11:55.4120378Z 2025-09-07T08:11:55.4120476Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4120978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4121460Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4121866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4122258Z layer_outputs = layer( 2025-09-07T08:11:55.4122632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4123028Z attn_outputs = self.attention( 2025-09-07T08:11:55.4123415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4123876Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4124315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:11:55.4124753Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:11:55.4124926Z 2025-09-07T08:11:55.4125020Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4125517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4126005Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4126424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4126817Z layer_outputs = layer( 2025-09-07T08:11:55.4127186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4127641Z attn_outputs = self.attention( 2025-09-07T08:11:55.4128030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4128483Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4128894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1324, in forward 2025-09-07T08:11:55.4129433Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:11:55.4130014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:11:55.4130527Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:11:55.4130723Z 2025-09-07T08:11:55.4130817Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4131319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4131804Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4132211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4132607Z layer_outputs = layer( 2025-09-07T08:11:55.4132970Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4133366Z attn_outputs = self.attention( 2025-09-07T08:11:55.4133754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T08:11:55.4134223Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T08:11:55.4134694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T08:11:55.4135100Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4135247Z 2025-09-07T08:11:55.4135342Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4135838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4136325Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4136730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4137123Z layer_outputs = layer( 2025-09-07T08:11:55.4137494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.4138084Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.4138576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.4138975Z return apply_chunking_to_forward( 2025-09-07T08:11:55.4139372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.4139767Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.4140180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:11:55.4140612Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4141010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T08:11:55.4141420Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4141562Z 2025-09-07T08:11:55.4141698Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4142197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4142735Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4143138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4143534Z layer_outputs = layer( 2025-09-07T08:11:55.4143909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.4144357Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.4144801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.4145207Z return apply_chunking_to_forward( 2025-09-07T08:11:55.4145598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.4145990Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.4146404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:11:55.4146827Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4147231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T08:11:55.4147642Z hidden_states = self.act_fn(hidden_states) 2025-09-07T08:11:55.4147780Z 2025-09-07T08:11:55.4147875Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4148377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4148867Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4149270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4149670Z layer_outputs = layer( 2025-09-07T08:11:55.4150044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.4150494Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.4150926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.4151341Z return apply_chunking_to_forward( 2025-09-07T08:11:55.4151726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.4152161Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.4152610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T08:11:55.4153018Z return self.output(hidden_states) 2025-09-07T08:11:55.4153422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T08:11:55.4153820Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4153951Z 2025-09-07T08:11:55.4154044Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4154538Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4155011Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4155412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4155824Z layer_outputs = layer( 2025-09-07T08:11:55.4156221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4156615Z attn_outputs = self.attention( 2025-09-07T08:11:55.4157053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4157480Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4157896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 540, in forward 2025-09-07T08:11:55.4158316Z query_key_vectors = self.query_key(hidden_states) 2025-09-07T08:11:55.4158463Z 2025-09-07T08:11:55.4158562Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4159056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4159539Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4159940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4160334Z layer_outputs = layer( 2025-09-07T08:11:55.4160696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4161089Z attn_outputs = self.attention( 2025-09-07T08:11:55.4161473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4161880Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4162288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:11:55.4162760Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:11:55.4163257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 749, in _hash_vectors 2025-09-07T08:11:55.4163792Z random_rotations = torch.randn(rotations_shape, device=vectors.device, dtype=vectors.dtype) 2025-09-07T08:11:55.4164042Z 2025-09-07T08:11:55.4164135Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4164624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4172957Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4173416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4173846Z layer_outputs = layer( 2025-09-07T08:11:55.4174326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4174776Z attn_outputs = self.attention( 2025-09-07T08:11:55.4175174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4175599Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4176023Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:11:55.4176508Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:11:55.4176993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 751, in _hash_vectors 2025-09-07T08:11:55.4177488Z rotated_vectors = torch.einsum("bmtd,mdhr->bmhtr", vectors, random_rotations) 2025-09-07T08:11:55.4177776Z 2025-09-07T08:11:55.4177882Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4178446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4178948Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4179417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4179822Z layer_outputs = layer( 2025-09-07T08:11:55.4180199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4180602Z attn_outputs = self.attention( 2025-09-07T08:11:55.4180990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4181401Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4181813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:11:55.4182277Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:11:55.4182758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T08:11:55.4183237Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T08:11:55.4183434Z 2025-09-07T08:11:55.4183534Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4184042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4184530Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4184935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4185334Z layer_outputs = layer( 2025-09-07T08:11:55.4185701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4186098Z attn_outputs = self.attention( 2025-09-07T08:11:55.4186482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4186891Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4187298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:11:55.4187767Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:11:55.4188251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T08:11:55.4188777Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T08:11:55.4188986Z 2025-09-07T08:11:55.4189086Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4189590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4190081Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4190494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4190890Z layer_outputs = layer( 2025-09-07T08:11:55.4191265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4191664Z attn_outputs = self.attention( 2025-09-07T08:11:55.4192053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4192511Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4192925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:11:55.4193483Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:11:55.4194096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:11:55.4194622Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:11:55.4195070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T08:11:55.4195573Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T08:11:55.4195782Z 2025-09-07T08:11:55.4195867Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4196095Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4196594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4197086Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4197501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4197897Z layer_outputs = layer( 2025-09-07T08:11:55.4198265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4198675Z attn_outputs = self.attention( 2025-09-07T08:11:55.4199066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4199482Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4199895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:11:55.4200413Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:11:55.4201016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:11:55.4201526Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:11:55.4201963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T08:11:55.4202396Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T08:11:55.4202535Z 2025-09-07T08:11:55.4202610Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4202848Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4203280Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4203774Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4204260Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4204664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4205057Z layer_outputs = layer( 2025-09-07T08:11:55.4205425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4205821Z attn_outputs = self.attention( 2025-09-07T08:11:55.4206196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4206609Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4207104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4207534Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4207999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:11:55.4208544Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4209096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:11:55.4209588Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:11:55.4209789Z 2025-09-07T08:11:55.4209872Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4210075Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4210298Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4210803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4211296Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4211716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4212113Z layer_outputs = layer( 2025-09-07T08:11:55.4212480Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4212878Z attn_outputs = self.attention( 2025-09-07T08:11:55.4213267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4213682Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4214094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4214521Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4214947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:11:55.4215469Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4216011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:11:55.4216421Z return torch.cat(slices, dim=3) 2025-09-07T08:11:55.4216545Z 2025-09-07T08:11:55.4216617Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4216834Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4217387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4217983Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4218392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4218782Z layer_outputs = layer( 2025-09-07T08:11:55.4219145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4219534Z attn_outputs = self.attention( 2025-09-07T08:11:55.4219914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4220321Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4220723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 541, in forward 2025-09-07T08:11:55.4221164Z value_vectors = self.value(hidden_states) 2025-09-07T08:11:55.4221295Z 2025-09-07T08:11:55.4221390Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4221987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4222469Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4222868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4223257Z layer_outputs = layer( 2025-09-07T08:11:55.4223617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4224010Z attn_outputs = self.attention( 2025-09-07T08:11:55.4224386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4224790Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4225189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4225603Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4226011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:11:55.4226511Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4227028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:11:55.4227506Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:11:55.4227706Z 2025-09-07T08:11:55.4227786Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4228003Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4228500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4228986Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4229399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4229794Z layer_outputs = layer( 2025-09-07T08:11:55.4230157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4230554Z attn_outputs = self.attention( 2025-09-07T08:11:55.4230941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4231438Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4231849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4232269Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4232686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:11:55.4233192Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4233717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:11:55.4234139Z return torch.cat(slices, dim=3) 2025-09-07T08:11:55.4234265Z 2025-09-07T08:11:55.4234364Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4234895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4235389Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4235828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4236222Z layer_outputs = layer( 2025-09-07T08:11:55.4236584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4236976Z attn_outputs = self.attention( 2025-09-07T08:11:55.4237352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4237758Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4238155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T08:11:55.4238681Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T08:11:55.4239247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T08:11:55.4239703Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T08:11:55.4239844Z 2025-09-07T08:11:55.4239942Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4240434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4240911Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4241310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4241709Z layer_outputs = layer( 2025-09-07T08:11:55.4242078Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4242473Z attn_outputs = self.attention( 2025-09-07T08:11:55.4242848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4243254Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4243655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4244075Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4244487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T08:11:55.4244953Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:11:55.4245255Z 2025-09-07T08:11:55.4245351Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4245879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4246365Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4246771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4247158Z layer_outputs = layer( 2025-09-07T08:11:55.4247530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4247925Z attn_outputs = self.attention( 2025-09-07T08:11:55.4248308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4248714Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4249166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4249589Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4250350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T08:11:55.4250875Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4251405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:11:55.4251891Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:11:55.4252100Z 2025-09-07T08:11:55.4252174Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4252272Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4252616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4252696Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4252968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4253027Z layer_outputs = layer( 2025-09-07T08:11:55.4253295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4253360Z attn_outputs = self.attention( 2025-09-07T08:11:55.4253623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4253704Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4253968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4254065Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4254333Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:11:55.4254444Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:11:55.4254451Z 2025-09-07T08:11:55.4254546Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4254885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4254970Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4255235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4255345Z layer_outputs = layer( 2025-09-07T08:11:55.4255650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4255718Z attn_outputs = self.attention( 2025-09-07T08:11:55.4255996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4256080Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4256351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4256444Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4256715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:11:55.4256818Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:11:55.4256823Z 2025-09-07T08:11:55.4256895Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4257023Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4257365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4257478Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4257797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4257859Z layer_outputs = layer( 2025-09-07T08:11:55.4258129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4258196Z attn_outputs = self.attention( 2025-09-07T08:11:55.4258468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4258552Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4258822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T08:11:55.4259025Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:11:55.4259336Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:11:55.4259473Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:11:55.4259476Z 2025-09-07T08:11:55.4259566Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4259905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4259986Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4260260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4260322Z layer_outputs = layer( 2025-09-07T08:11:55.4260592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4260661Z attn_outputs = self.attention( 2025-09-07T08:11:55.4260927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T08:11:55.4261063Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T08:11:55.4261326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T08:11:55.4261403Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4261410Z 2025-09-07T08:11:55.4261537Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4261917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4261999Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4262266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4262329Z layer_outputs = layer( 2025-09-07T08:11:55.4262593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1601, in forward 2025-09-07T08:11:55.4262675Z attn_output = prev_attn_output + attn_output 2025-09-07T08:11:55.4262678Z 2025-09-07T08:11:55.4262769Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4263103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4263219Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4263487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4263581Z layer_outputs = layer( 2025-09-07T08:11:55.4263851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.4263970Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.4264242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.4264313Z return apply_chunking_to_forward( 2025-09-07T08:11:55.4264585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.4264654Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.4264950Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:11:55.4265033Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4265302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T08:11:55.4265382Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4265385Z 2025-09-07T08:11:55.4265478Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4265822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4265903Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4266168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4266238Z layer_outputs = layer( 2025-09-07T08:11:55.4266505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.4266626Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.4266899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.4266966Z return apply_chunking_to_forward( 2025-09-07T08:11:55.4267224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.4267291Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.4267578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:11:55.4267692Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4267998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T08:11:55.4268074Z hidden_states = self.act_fn(hidden_states) 2025-09-07T08:11:55.4268077Z 2025-09-07T08:11:55.4268170Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4268518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4268597Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4268871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4268933Z layer_outputs = layer( 2025-09-07T08:11:55.4269196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.4269343Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.4269613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.4269684Z return apply_chunking_to_forward( 2025-09-07T08:11:55.4269985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.4270054Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.4270338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T08:11:55.4270405Z return self.output(hidden_states) 2025-09-07T08:11:55.4270676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T08:11:55.4270749Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4270754Z 2025-09-07T08:11:55.4270850Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4271187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4271273Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4271540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4271598Z layer_outputs = layer( 2025-09-07T08:11:55.4271868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.4271978Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.4271981Z 2025-09-07T08:11:55.4272078Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4272415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4272492Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4272758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4272816Z layer_outputs = layer( 2025-09-07T08:11:55.4273081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4273147Z attn_outputs = self.attention( 2025-09-07T08:11:55.4273412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4273493Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4273762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1198, in forward 2025-09-07T08:11:55.4273908Z query_vectors = self.query(hidden_states) 2025-09-07T08:11:55.4273911Z 2025-09-07T08:11:55.4274000Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4274342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4274421Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4274692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4274752Z layer_outputs = layer( 2025-09-07T08:11:55.4275016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4275088Z attn_outputs = self.attention( 2025-09-07T08:11:55.4275354Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4275495Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4275765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1199, in forward 2025-09-07T08:11:55.4275869Z key_vectors = self.key(hidden_states) 2025-09-07T08:11:55.4275873Z 2025-09-07T08:11:55.4275954Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4276026Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4276117Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4276455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4276532Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4276800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4276862Z layer_outputs = layer( 2025-09-07T08:11:55.4277132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4277195Z attn_outputs = self.attention( 2025-09-07T08:11:55.4277462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4277540Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4277804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:11:55.4277987Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4278271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:11:55.4278418Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:11:55.4278422Z 2025-09-07T08:11:55.4278493Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4278568Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4278662Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4279000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4279079Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4279346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4279407Z layer_outputs = layer( 2025-09-07T08:11:55.4279671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4279776Z attn_outputs = self.attention( 2025-09-07T08:11:55.4280080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4280159Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4280427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T08:11:55.4280594Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4280875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:11:55.4280943Z return torch.cat(slices, dim=3) 2025-09-07T08:11:55.4280947Z 2025-09-07T08:11:55.4281035Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4281408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4281490Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4281792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4281851Z layer_outputs = layer( 2025-09-07T08:11:55.4282115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4282181Z attn_outputs = self.attention( 2025-09-07T08:11:55.4282446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4282525Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4282788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:11:55.4282935Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:11:55.4282939Z 2025-09-07T08:11:55.4283027Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4283362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4283440Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4283706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4283767Z layer_outputs = layer( 2025-09-07T08:11:55.4284029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4284094Z attn_outputs = self.attention( 2025-09-07T08:11:55.4284358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4284437Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4284709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T08:11:55.4284846Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:11:55.4284849Z 2025-09-07T08:11:55.4284943Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4285280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4285355Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4285628Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4285724Z layer_outputs = layer( 2025-09-07T08:11:55.4286030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4286095Z attn_outputs = self.attention( 2025-09-07T08:11:55.4286364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4286440Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4286707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1200, in forward 2025-09-07T08:11:55.4286781Z value_vectors = self.value(hidden_states) 2025-09-07T08:11:55.4286784Z 2025-09-07T08:11:55.4286853Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4286924Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4287012Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4287381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4287462Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4287758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4287819Z layer_outputs = layer( 2025-09-07T08:11:55.4288085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4288151Z attn_outputs = self.attention( 2025-09-07T08:11:55.4288414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4288490Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4288754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1266, in forward 2025-09-07T08:11:55.4288943Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4289228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:11:55.4289363Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:11:55.4289367Z 2025-09-07T08:11:55.4289438Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4289507Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4289596Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4289936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4290010Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4290280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4290338Z layer_outputs = layer( 2025-09-07T08:11:55.4290601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4290666Z attn_outputs = self.attention( 2025-09-07T08:11:55.4290931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4291009Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4291277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:11:55.4291387Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:11:55.4291393Z 2025-09-07T08:11:55.4291491Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4291876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4291994Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4292272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4292337Z layer_outputs = layer( 2025-09-07T08:11:55.4292604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4292670Z attn_outputs = self.attention( 2025-09-07T08:11:55.4292937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4293021Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4293287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T08:11:55.4293439Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:11:55.4293442Z 2025-09-07T08:11:55.4293538Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4293911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4293992Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4294261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4294322Z layer_outputs = layer( 2025-09-07T08:11:55.4294590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4294652Z attn_outputs = self.attention( 2025-09-07T08:11:55.4294920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4295003Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4295268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1324, in forward 2025-09-07T08:11:55.4295474Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:11:55.4295788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:11:55.4295926Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:11:55.4295929Z 2025-09-07T08:11:55.4296022Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4296366Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4296458Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4296726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4296794Z layer_outputs = layer( 2025-09-07T08:11:55.4297062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4297131Z attn_outputs = self.attention( 2025-09-07T08:11:55.4297396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T08:11:55.4297530Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T08:11:55.4297862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T08:11:55.4297977Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4298010Z 2025-09-07T08:11:55.4298105Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4298443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4298522Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4298786Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4298844Z layer_outputs = layer( 2025-09-07T08:11:55.4299111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.4299228Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.4299495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.4299601Z return apply_chunking_to_forward( 2025-09-07T08:11:55.4299868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.4299968Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.4300259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:11:55.4300334Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4300612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T08:11:55.4300688Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4300692Z 2025-09-07T08:11:55.4300782Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4301125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4301209Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4301476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4301536Z layer_outputs = layer( 2025-09-07T08:11:55.4301799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.4301906Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.4302175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.4302249Z return apply_chunking_to_forward( 2025-09-07T08:11:55.4302516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.4302587Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.4302881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:11:55.4302957Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4303374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T08:11:55.4303462Z hidden_states = self.act_fn(hidden_states) 2025-09-07T08:11:55.4303465Z 2025-09-07T08:11:55.4303560Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4303908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4303988Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4304363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4304469Z layer_outputs = layer( 2025-09-07T08:11:55.4304737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.4304863Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.4305130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.4305205Z return apply_chunking_to_forward( 2025-09-07T08:11:55.4305464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.4305531Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.4305825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T08:11:55.4305897Z return self.output(hidden_states) 2025-09-07T08:11:55.4306214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T08:11:55.4306291Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4306360Z 2025-09-07T08:11:55.4306461Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4306803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4306881Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4307157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4307219Z layer_outputs = layer( 2025-09-07T08:11:55.4307493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4307566Z attn_outputs = self.attention( 2025-09-07T08:11:55.4307836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4307929Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4308201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 540, in forward 2025-09-07T08:11:55.4308293Z query_key_vectors = self.query_key(hidden_states) 2025-09-07T08:11:55.4308297Z 2025-09-07T08:11:55.4308390Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4308735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4308812Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4309081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4309148Z layer_outputs = layer( 2025-09-07T08:11:55.4309417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4309491Z attn_outputs = self.attention( 2025-09-07T08:11:55.4309757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4309837Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4310110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:11:55.4310250Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:11:55.4310539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 749, in _hash_vectors 2025-09-07T08:11:55.4310793Z random_rotations = torch.randn(rotations_shape, device=vectors.device, dtype=vectors.dtype) 2025-09-07T08:11:55.4310797Z 2025-09-07T08:11:55.4310893Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4311235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4311316Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4311591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4311652Z layer_outputs = layer( 2025-09-07T08:11:55.4311923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4311989Z attn_outputs = self.attention( 2025-09-07T08:11:55.4312295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4312382Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4312678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:11:55.4312817Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:11:55.4313100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 751, in _hash_vectors 2025-09-07T08:11:55.4313254Z rotated_vectors = torch.einsum("bmtd,mdhr->bmhtr", vectors, random_rotations) 2025-09-07T08:11:55.4313257Z 2025-09-07T08:11:55.4313350Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4313708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4313791Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4314055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4314125Z layer_outputs = layer( 2025-09-07T08:11:55.4314394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4314466Z attn_outputs = self.attention( 2025-09-07T08:11:55.4314734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4314814Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4315088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:11:55.4315222Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:11:55.4315509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T08:11:55.4315644Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T08:11:55.4315647Z 2025-09-07T08:11:55.4315746Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4316092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4316172Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4316450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4316513Z layer_outputs = layer( 2025-09-07T08:11:55.4316787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4316936Z attn_outputs = self.attention( 2025-09-07T08:11:55.4317204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4317296Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4317565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T08:11:55.4317708Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T08:11:55.4317994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T08:11:55.4318089Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T08:11:55.4318092Z 2025-09-07T08:11:55.4318187Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4318560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4318649Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4318949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4319018Z layer_outputs = layer( 2025-09-07T08:11:55.4319282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4319353Z attn_outputs = self.attention( 2025-09-07T08:11:55.4319621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4319702Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4319973Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:11:55.4320167Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:11:55.4320541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:11:55.4320642Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:11:55.4320939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T08:11:55.4321087Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T08:11:55.4321090Z 2025-09-07T08:11:55.4321166Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4321271Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4321614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4321702Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4321971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4322036Z layer_outputs = layer( 2025-09-07T08:11:55.4322307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4322373Z attn_outputs = self.attention( 2025-09-07T08:11:55.4322648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4322728Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4322999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T08:11:55.4323248Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T08:11:55.4323613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T08:11:55.4323714Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T08:11:55.4324001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T08:11:55.4324092Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T08:11:55.4324095Z 2025-09-07T08:11:55.4324165Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4324243Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4324336Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4324708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4324798Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4325106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4325177Z layer_outputs = layer( 2025-09-07T08:11:55.4325445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4325510Z attn_outputs = self.attention( 2025-09-07T08:11:55.4325784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4325863Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4326136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4326234Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4326501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:11:55.4326713Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4326999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:11:55.4327151Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:11:55.4327155Z 2025-09-07T08:11:55.4327227Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4327305Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4327398Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4327740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4327833Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4328102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4328170Z layer_outputs = layer( 2025-09-07T08:11:55.4328436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4328506Z attn_outputs = self.attention( 2025-09-07T08:11:55.4328773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4328853Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4329123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4329276Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4329548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T08:11:55.4329743Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4330029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:11:55.4330107Z return torch.cat(slices, dim=3) 2025-09-07T08:11:55.4330110Z 2025-09-07T08:11:55.4330182Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4330282Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4330619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4330707Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4331012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4331077Z layer_outputs = layer( 2025-09-07T08:11:55.4331384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4331453Z attn_outputs = self.attention( 2025-09-07T08:11:55.4331727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4331809Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4332075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 541, in forward 2025-09-07T08:11:55.4332158Z value_vectors = self.value(hidden_states) 2025-09-07T08:11:55.4332167Z 2025-09-07T08:11:55.4332261Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4332611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4332690Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4332963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4333022Z layer_outputs = layer( 2025-09-07T08:11:55.4333288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4333359Z attn_outputs = self.attention( 2025-09-07T08:11:55.4333622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4333704Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4333968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4334061Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4334324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:11:55.4334499Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4334782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:11:55.4334913Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:11:55.4334917Z 2025-09-07T08:11:55.4334993Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4335084Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4335458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4335569Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4335840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4335904Z layer_outputs = layer( 2025-09-07T08:11:55.4336171Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4336238Z attn_outputs = self.attention( 2025-09-07T08:11:55.4336503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4336581Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4336848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4337029Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4337298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T08:11:55.4337495Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4337826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T08:11:55.4337894Z return torch.cat(slices, dim=3) 2025-09-07T08:11:55.4337898Z 2025-09-07T08:11:55.4337988Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4338328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4338406Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4338680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4338739Z layer_outputs = layer( 2025-09-07T08:11:55.4339003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4339067Z attn_outputs = self.attention( 2025-09-07T08:11:55.4339333Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4339413Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4339682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T08:11:55.4339885Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T08:11:55.4340195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T08:11:55.4340280Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T08:11:55.4340283Z 2025-09-07T08:11:55.4340374Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4340709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4340792Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4341056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4341116Z layer_outputs = layer( 2025-09-07T08:11:55.4341382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4341501Z attn_outputs = self.attention( 2025-09-07T08:11:55.4341806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4341883Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4342152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4342240Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4342500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T08:11:55.4342646Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T08:11:55.4342649Z 2025-09-07T08:11:55.4342741Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4343116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4343197Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4343466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4343555Z layer_outputs = layer( 2025-09-07T08:11:55.4343826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4343896Z attn_outputs = self.attention( 2025-09-07T08:11:55.4344164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4344244Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4344509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4344598Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4344868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T08:11:55.4345056Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T08:11:55.4345347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T08:11:55.4345481Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T08:11:55.4345491Z 2025-09-07T08:11:55.4345563Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4345653Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4345991Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4346072Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4346343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4346403Z layer_outputs = layer( 2025-09-07T08:11:55.4346667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4346731Z attn_outputs = self.attention( 2025-09-07T08:11:55.4346998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4347075Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4347349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4347433Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4347744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:11:55.4347885Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:11:55.4347888Z 2025-09-07T08:11:55.4347979Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4348321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4348399Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4348665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4348724Z layer_outputs = layer( 2025-09-07T08:11:55.4348996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4349064Z attn_outputs = self.attention( 2025-09-07T08:11:55.4349375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4349465Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4349772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T08:11:55.4349873Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T08:11:55.4350142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T08:11:55.4350250Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T08:11:55.4350259Z 2025-09-07T08:11:55.4350333Z cudagraph partition due to non gpu ops 2025-09-07T08:11:55.4350430Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4350782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4350869Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4351143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4351208Z layer_outputs = layer( 2025-09-07T08:11:55.4351473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4351544Z attn_outputs = self.attention( 2025-09-07T08:11:55.4351809Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T08:11:55.4351893Z self_attention_outputs = self.self_attention( 2025-09-07T08:11:55.4352161Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T08:11:55.4352375Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T08:11:55.4352697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T08:11:55.4352835Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T08:11:55.4352838Z 2025-09-07T08:11:55.4352938Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4353280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4353361Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4353627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4353726Z layer_outputs = layer( 2025-09-07T08:11:55.4354028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T08:11:55.4354093Z attn_outputs = self.attention( 2025-09-07T08:11:55.4354361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T08:11:55.4354494Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T08:11:55.4354759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T08:11:55.4354836Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4354840Z 2025-09-07T08:11:55.4354929Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4355268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4355381Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4355652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4355741Z layer_outputs = layer( 2025-09-07T08:11:55.4356011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.4356129Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.4356392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.4356465Z return apply_chunking_to_forward( 2025-09-07T08:11:55.4356728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.4356802Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.4357092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:11:55.4357165Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4357433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T08:11:55.4357502Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4357505Z 2025-09-07T08:11:55.4357596Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4357933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4358011Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4358274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4358334Z layer_outputs = layer( 2025-09-07T08:11:55.4358606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.4358717Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.4358986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.4359057Z return apply_chunking_to_forward( 2025-09-07T08:11:55.4359311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.4359379Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.4359660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T08:11:55.4359767Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4360068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T08:11:55.4360146Z hidden_states = self.act_fn(hidden_states) 2025-09-07T08:11:55.4360149Z 2025-09-07T08:11:55.4360242Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4360580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4360658Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4360921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4360987Z layer_outputs = layer( 2025-09-07T08:11:55.4361254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.4361405Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.4361678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T08:11:55.4361749Z return apply_chunking_to_forward( 2025-09-07T08:11:55.4362065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:55.4362136Z return forward_fn(*input_tensors) 2025-09-07T08:11:55.4362429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T08:11:55.4362500Z return self.output(hidden_states) 2025-09-07T08:11:55.4362769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T08:11:55.4362850Z hidden_states = self.dense(hidden_states) 2025-09-07T08:11:55.4362856Z 2025-09-07T08:11:55.4362949Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4363297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4363377Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4363649Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T08:11:55.4363710Z layer_outputs = layer( 2025-09-07T08:11:55.4363976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T08:11:55.4364094Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T08:11:55.4364097Z 2025-09-07T08:11:55.4364189Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:55.4364535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T08:11:55.4364613Z hidden_states = _ReversibleFunction.apply( 2025-09-07T08:11:55.4364887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1753, in forward 2025-09-07T08:11:55.4364992Z return torch.cat([attn_output, hidden_states], dim=-1) 2025-09-07T08:11:55.4364995Z 2025-09-07T08:11:58.5754051Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:11:58.5754698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2495, in torch_dynamo_resume_in_forward_at_2481 2025-09-07T08:11:58.5755232Z logits = self.lm_head(sequence_output) 2025-09-07T08:11:58.5755677Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1886, in forward 2025-09-07T08:11:58.5756609Z return apply_chunking_to_forward(self.forward_chunk, self.chunk_size_lm_head, self.seq_len_dim, hidden_states) 2025-09-07T08:11:58.5757277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:11:58.5757702Z return forward_fn(*input_tensors) 2025-09-07T08:11:58.5758131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1889, in forward_chunk 2025-09-07T08:11:58.5758572Z hidden_states = self.decoder(hidden_states) 2025-09-07T08:11:58.5758718Z 2025-09-07T08:11:59.4170210Z pass 2025-09-07T08:11:59.4170591Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:12:01.7739385Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:12:01.7740775Z import pynvml # type: ignore[import] 2025-09-07T08:12:03.7821064Z 2025-09-07T08:12:09.5487796Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:12:09.5488079Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:12:09.5826851Z cpu eval hf_Roberta_base 2025-09-07T08:12:11.5229556Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:12:12.0577377Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:12:12.5948140Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:12:20.0487832Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0488281Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0488665Z return mod(**inputs) 2025-09-07T08:12:20.0489131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0489567Z outputs = self.roberta( 2025-09-07T08:12:20.0489992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 817, in forward 2025-09-07T08:12:20.0490432Z embedding_output = self.embeddings( 2025-09-07T08:12:20.0490852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 90, in forward 2025-09-07T08:12:20.0491428Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx, past_key_values_length) 2025-09-07T08:12:20.0492051Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1576, in create_position_ids_from_input_ids 2025-09-07T08:12:20.0492543Z mask = input_ids.ne(padding_idx).int() 2025-09-07T08:12:20.0492679Z 2025-09-07T08:12:20.0492769Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0492972Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0493182Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0493373Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0493566Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0493747Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0493968Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0494166Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0494365Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0494548Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0494737Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0494931Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0495160Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0495530Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0496282Z return mod(**inputs) 2025-09-07T08:12:20.0496683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0497100Z outputs = self.roberta( 2025-09-07T08:12:20.0497490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 817, in forward 2025-09-07T08:12:20.0497963Z embedding_output = self.embeddings( 2025-09-07T08:12:20.0498371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 90, in forward 2025-09-07T08:12:20.0498911Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx, past_key_values_length) 2025-09-07T08:12:20.0499515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1577, in create_position_ids_from_input_ids 2025-09-07T08:12:20.0500190Z incremental_indices = (torch.cumsum(mask, dim=1).type_as(mask) + past_key_values_length) * mask 2025-09-07T08:12:20.0500443Z 2025-09-07T08:12:20.0500552Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0500966Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0501291Z return mod(**inputs) 2025-09-07T08:12:20.0501693Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0502106Z outputs = self.roberta( 2025-09-07T08:12:20.0502478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 817, in forward 2025-09-07T08:12:20.0502899Z embedding_output = self.embeddings( 2025-09-07T08:12:20.0503453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 90, in forward 2025-09-07T08:12:20.0504026Z position_ids = create_position_ids_from_input_ids(input_ids, self.padding_idx, past_key_values_length) 2025-09-07T08:12:20.0504662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1577, in create_position_ids_from_input_ids 2025-09-07T08:12:20.0505252Z incremental_indices = (torch.cumsum(mask, dim=1).type_as(mask) + past_key_values_length) * mask 2025-09-07T08:12:20.0505500Z 2025-09-07T08:12:20.0505606Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0505960Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0506278Z return mod(**inputs) 2025-09-07T08:12:20.0506663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0507071Z outputs = self.roberta( 2025-09-07T08:12:20.0507466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 817, in forward 2025-09-07T08:12:20.0507881Z embedding_output = self.embeddings( 2025-09-07T08:12:20.0508283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 120, in forward 2025-09-07T08:12:20.0508706Z embeddings = self.LayerNorm(embeddings) 2025-09-07T08:12:20.0508839Z 2025-09-07T08:12:20.0508932Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0509282Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0509600Z return mod(**inputs) 2025-09-07T08:12:20.0509978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0510415Z outputs = self.roberta( 2025-09-07T08:12:20.0510850Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0511308Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0511713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0512123Z layer_outputs = layer_module( 2025-09-07T08:12:20.0512459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0512830Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0513245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0513659Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0514036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0514409Z return func(*args, **kwargs) 2025-09-07T08:12:20.0514866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0515278Z self_outputs = self.self( 2025-09-07T08:12:20.0515684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0516075Z return func(*args, **kwargs) 2025-09-07T08:12:20.0516496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T08:12:20.0517072Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:12:20.0517347Z 2025-09-07T08:12:20.0517447Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0517814Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0518135Z return mod(**inputs) 2025-09-07T08:12:20.0518521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0518932Z outputs = self.roberta( 2025-09-07T08:12:20.0519316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0519728Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0520130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0520536Z layer_outputs = layer_module( 2025-09-07T08:12:20.0520882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0521246Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0521659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0522083Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0522473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0522848Z return func(*args, **kwargs) 2025-09-07T08:12:20.0523241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0523651Z self_outputs = self.self( 2025-09-07T08:12:20.0523997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0524360Z return func(*args, **kwargs) 2025-09-07T08:12:20.0524748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T08:12:20.0525230Z self.key(current_states) 2025-09-07T08:12:20.0525345Z 2025-09-07T08:12:20.0525448Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0525798Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0526113Z return mod(**inputs) 2025-09-07T08:12:20.0526516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0526919Z outputs = self.roberta( 2025-09-07T08:12:20.0527300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0527708Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0528093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0528496Z layer_outputs = layer_module( 2025-09-07T08:12:20.0528895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0529263Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0529696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0530116Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0530488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0530856Z return func(*args, **kwargs) 2025-09-07T08:12:20.0531245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0531642Z self_outputs = self.self( 2025-09-07T08:12:20.0531992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0532365Z return func(*args, **kwargs) 2025-09-07T08:12:20.0532746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T08:12:20.0533145Z self.value(current_states) 2025-09-07T08:12:20.0533258Z 2025-09-07T08:12:20.0533336Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0533565Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0533917Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0534234Z return mod(**inputs) 2025-09-07T08:12:20.0534621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0535031Z outputs = self.roberta( 2025-09-07T08:12:20.0535411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0535821Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0536215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0536608Z layer_outputs = layer_module( 2025-09-07T08:12:20.0536940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0537299Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0537766Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0538203Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0538592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0539020Z return func(*args, **kwargs) 2025-09-07T08:12:20.0539453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0549384Z self_outputs = self.self( 2025-09-07T08:12:20.0549823Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0550227Z return func(*args, **kwargs) 2025-09-07T08:12:20.0550642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:12:20.0551128Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:12:20.0551315Z 2025-09-07T08:12:20.0551424Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0551790Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0552116Z return mod(**inputs) 2025-09-07T08:12:20.0552583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0553008Z outputs = self.roberta( 2025-09-07T08:12:20.0553430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0553843Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0554239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0554640Z layer_outputs = layer_module( 2025-09-07T08:12:20.0554977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0555338Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0555746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0556171Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0556548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0556914Z return func(*args, **kwargs) 2025-09-07T08:12:20.0557302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T08:12:20.0557757Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:12:20.0558213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T08:12:20.0558629Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0558763Z 2025-09-07T08:12:20.0558863Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0559228Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0559551Z return mod(**inputs) 2025-09-07T08:12:20.0559937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0560344Z outputs = self.roberta( 2025-09-07T08:12:20.0560718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0561125Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0561525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0561935Z layer_outputs = layer_module( 2025-09-07T08:12:20.0562266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0562621Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0563077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0563531Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0563934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0564323Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0564763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0565250Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0565700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T08:12:20.0566117Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0566247Z 2025-09-07T08:12:20.0566346Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0566733Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0567059Z return mod(**inputs) 2025-09-07T08:12:20.0567469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0567868Z outputs = self.roberta( 2025-09-07T08:12:20.0568250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0568669Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0569092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0569503Z layer_outputs = layer_module( 2025-09-07T08:12:20.0569834Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0570195Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0570601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0571018Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0571418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0571804Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0572239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0572718Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0573169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T08:12:20.0573609Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:12:20.0573985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:12:20.0574318Z return self.act(input) 2025-09-07T08:12:20.0574430Z 2025-09-07T08:12:20.0574532Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0574887Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0575200Z return mod(**inputs) 2025-09-07T08:12:20.0575581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0575988Z outputs = self.roberta( 2025-09-07T08:12:20.0576372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0576775Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0577207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0577724Z layer_outputs = layer_module( 2025-09-07T08:12:20.0578062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0578422Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0578848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0579255Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0579648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0580038Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0580477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T08:12:20.0581013Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:12:20.0581473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T08:12:20.0581940Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0582082Z 2025-09-07T08:12:20.0582341Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0582698Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0583017Z return mod(**inputs) 2025-09-07T08:12:20.0583390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0583798Z outputs = self.roberta( 2025-09-07T08:12:20.0584182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0584599Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0584986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0585392Z layer_outputs = layer_module( 2025-09-07T08:12:20.0585729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0586084Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0586486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0586892Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0587270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0587655Z return func(*args, **kwargs) 2025-09-07T08:12:20.0588045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0588447Z self_outputs = self.self( 2025-09-07T08:12:20.0588797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0589161Z return func(*args, **kwargs) 2025-09-07T08:12:20.0589543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T08:12:20.0590088Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:12:20.0590348Z 2025-09-07T08:12:20.0590446Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0590798Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0591162Z return mod(**inputs) 2025-09-07T08:12:20.0591550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0591997Z outputs = self.roberta( 2025-09-07T08:12:20.0592374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0592781Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0593177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0593579Z layer_outputs = layer_module( 2025-09-07T08:12:20.0593912Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0594259Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0594663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0595110Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0595485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0595846Z return func(*args, **kwargs) 2025-09-07T08:12:20.0596264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0596675Z self_outputs = self.self( 2025-09-07T08:12:20.0597032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0597401Z return func(*args, **kwargs) 2025-09-07T08:12:20.0597785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T08:12:20.0598196Z self.key(current_states) 2025-09-07T08:12:20.0598313Z 2025-09-07T08:12:20.0598413Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0598771Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0599093Z return mod(**inputs) 2025-09-07T08:12:20.0599481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0599889Z outputs = self.roberta( 2025-09-07T08:12:20.0600268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0600674Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0601067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0601471Z layer_outputs = layer_module( 2025-09-07T08:12:20.0601814Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0602183Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0602587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0602992Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0603575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0603947Z return func(*args, **kwargs) 2025-09-07T08:12:20.0604335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0604742Z self_outputs = self.self( 2025-09-07T08:12:20.0605090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0605546Z return func(*args, **kwargs) 2025-09-07T08:12:20.0605983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T08:12:20.0606385Z self.value(current_states) 2025-09-07T08:12:20.0606495Z 2025-09-07T08:12:20.0606573Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0606807Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0607162Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0607482Z return mod(**inputs) 2025-09-07T08:12:20.0607864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0608268Z outputs = self.roberta( 2025-09-07T08:12:20.0608652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0609057Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0609501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0609907Z layer_outputs = layer_module( 2025-09-07T08:12:20.0611436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0611810Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0612220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0612633Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0612998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0613368Z return func(*args, **kwargs) 2025-09-07T08:12:20.0613758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0614168Z self_outputs = self.self( 2025-09-07T08:12:20.0614514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0614879Z return func(*args, **kwargs) 2025-09-07T08:12:20.0615262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:12:20.0615731Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:12:20.0615915Z 2025-09-07T08:12:20.0616019Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0616365Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0616692Z return mod(**inputs) 2025-09-07T08:12:20.0617075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0617484Z outputs = self.roberta( 2025-09-07T08:12:20.0617933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0618333Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0618732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0619134Z layer_outputs = layer_module( 2025-09-07T08:12:20.0619469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0619819Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0620224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0620637Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0621060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0621460Z return func(*args, **kwargs) 2025-09-07T08:12:20.0621839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T08:12:20.0622296Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:12:20.0622746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T08:12:20.0623165Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0623299Z 2025-09-07T08:12:20.0623401Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0623742Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0624062Z return mod(**inputs) 2025-09-07T08:12:20.0624475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0624888Z outputs = self.roberta( 2025-09-07T08:12:20.0625302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0625707Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0626098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0626500Z layer_outputs = layer_module( 2025-09-07T08:12:20.0626832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0627179Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0627580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0627998Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0628400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0628794Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0629224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0629710Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0630159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T08:12:20.0630571Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0630701Z 2025-09-07T08:12:20.0630803Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0631150Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0631471Z return mod(**inputs) 2025-09-07T08:12:20.0631848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0632255Z outputs = self.roberta( 2025-09-07T08:12:20.0632630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0633033Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0633423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0633819Z layer_outputs = layer_module( 2025-09-07T08:12:20.0634143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0634486Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0634940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0635384Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0635771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0636149Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0636566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0637032Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0637463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T08:12:20.0637897Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:12:20.0638303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:12:20.0638633Z return self.act(input) 2025-09-07T08:12:20.0638742Z 2025-09-07T08:12:20.0638834Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0639213Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0639522Z return mod(**inputs) 2025-09-07T08:12:20.0639895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0640294Z outputs = self.roberta( 2025-09-07T08:12:20.0640660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0641053Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0641434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0641828Z layer_outputs = layer_module( 2025-09-07T08:12:20.0642157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0642497Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0642894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0643297Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0643676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0644062Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0644482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T08:12:20.0644970Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:12:20.0645429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T08:12:20.0645850Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0645987Z 2025-09-07T08:12:20.0646084Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0646443Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0646762Z return mod(**inputs) 2025-09-07T08:12:20.0647139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0647548Z outputs = self.roberta( 2025-09-07T08:12:20.0647928Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0648388Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0648822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0649219Z layer_outputs = layer_module( 2025-09-07T08:12:20.0649561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0649921Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0650316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0650721Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0651085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0651453Z return func(*args, **kwargs) 2025-09-07T08:12:20.0651829Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0652268Z self_outputs = self.self( 2025-09-07T08:12:20.0652617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0652969Z return func(*args, **kwargs) 2025-09-07T08:12:20.0653427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T08:12:20.0653975Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:12:20.0654235Z 2025-09-07T08:12:20.0654331Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0654673Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0654984Z return mod(**inputs) 2025-09-07T08:12:20.0655357Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0655759Z outputs = self.roberta( 2025-09-07T08:12:20.0656132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0656536Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0656919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0657316Z layer_outputs = layer_module( 2025-09-07T08:12:20.0657700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0658055Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0658441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0658849Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0659223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0659584Z return func(*args, **kwargs) 2025-09-07T08:12:20.0659957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0660356Z self_outputs = self.self( 2025-09-07T08:12:20.0660690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0661055Z return func(*args, **kwargs) 2025-09-07T08:12:20.0661429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T08:12:20.0661819Z self.key(current_states) 2025-09-07T08:12:20.0661924Z 2025-09-07T08:12:20.0662019Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0662402Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0662752Z return mod(**inputs) 2025-09-07T08:12:20.0663118Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0663521Z outputs = self.roberta( 2025-09-07T08:12:20.0663899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0664300Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0664685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0665084Z layer_outputs = layer_module( 2025-09-07T08:12:20.0665408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0665763Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0666195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0666601Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0666990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0667360Z return func(*args, **kwargs) 2025-09-07T08:12:20.0667738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0668129Z self_outputs = self.self( 2025-09-07T08:12:20.0668468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0668818Z return func(*args, **kwargs) 2025-09-07T08:12:20.0669197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T08:12:20.0669607Z self.value(current_states) 2025-09-07T08:12:20.0669717Z 2025-09-07T08:12:20.0669793Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0670003Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0670348Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0670650Z return mod(**inputs) 2025-09-07T08:12:20.0671013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0671414Z outputs = self.roberta( 2025-09-07T08:12:20.0671780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0672174Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0672565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0672959Z layer_outputs = layer_module( 2025-09-07T08:12:20.0673277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0673620Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0674014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0674425Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0674781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0675134Z return func(*args, **kwargs) 2025-09-07T08:12:20.0675512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0675950Z self_outputs = self.self( 2025-09-07T08:12:20.0676315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0676679Z return func(*args, **kwargs) 2025-09-07T08:12:20.0677053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:12:20.0677501Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:12:20.0677680Z 2025-09-07T08:12:20.0677769Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0678104Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0678413Z return mod(**inputs) 2025-09-07T08:12:20.0678793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0679188Z outputs = self.roberta( 2025-09-07T08:12:20.0679587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0679989Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0680403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0680818Z layer_outputs = layer_module( 2025-09-07T08:12:20.0681153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0681510Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0681907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0682313Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0682681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0683050Z return func(*args, **kwargs) 2025-09-07T08:12:20.0683432Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T08:12:20.0683889Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:12:20.0684335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T08:12:20.0684749Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0684888Z 2025-09-07T08:12:20.0684981Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0685325Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0685629Z return mod(**inputs) 2025-09-07T08:12:20.0686007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0686420Z outputs = self.roberta( 2025-09-07T08:12:20.0686795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0687200Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0687582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0687980Z layer_outputs = layer_module( 2025-09-07T08:12:20.0688310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0688661Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0689059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0689517Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0689953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0690347Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0690783Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0691282Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0691733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T08:12:20.0692145Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0692280Z 2025-09-07T08:12:20.0692378Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0692724Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0693035Z return mod(**inputs) 2025-09-07T08:12:20.0693465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0693871Z outputs = self.roberta( 2025-09-07T08:12:20.0694281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0694686Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0695073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0695474Z layer_outputs = layer_module( 2025-09-07T08:12:20.0695811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0696163Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0696562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0696973Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0697370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0697813Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0698242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0698715Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0699158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T08:12:20.0699598Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:12:20.0699971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:12:20.0700316Z return self.act(input) 2025-09-07T08:12:20.0700421Z 2025-09-07T08:12:20.0700517Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0700870Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0701190Z return mod(**inputs) 2025-09-07T08:12:20.0701567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0701970Z outputs = self.roberta( 2025-09-07T08:12:20.0702340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0702745Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0703275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0703765Z layer_outputs = layer_module( 2025-09-07T08:12:20.0704162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0704513Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0704919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0705335Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0705734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0706122Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0706553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T08:12:20.0707053Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:12:20.0707600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T08:12:20.0708027Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0708161Z 2025-09-07T08:12:20.0708258Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0708658Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0708982Z return mod(**inputs) 2025-09-07T08:12:20.0709367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0709773Z outputs = self.roberta( 2025-09-07T08:12:20.0710150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0710550Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0710947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0711370Z layer_outputs = layer_module( 2025-09-07T08:12:20.0711712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0712065Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0712466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0712879Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0713253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0713617Z return func(*args, **kwargs) 2025-09-07T08:12:20.0714013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0714424Z self_outputs = self.self( 2025-09-07T08:12:20.0714785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0715156Z return func(*args, **kwargs) 2025-09-07T08:12:20.0715538Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T08:12:20.0716088Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:12:20.0716359Z 2025-09-07T08:12:20.0716457Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0716814Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0717134Z return mod(**inputs) 2025-09-07T08:12:20.0717508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0717965Z outputs = self.roberta( 2025-09-07T08:12:20.0718410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0718819Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0719217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0719619Z layer_outputs = layer_module( 2025-09-07T08:12:20.0719965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0720326Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0720733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0721141Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0721522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0721933Z return func(*args, **kwargs) 2025-09-07T08:12:20.0722323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0722756Z self_outputs = self.self( 2025-09-07T08:12:20.0723102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0723464Z return func(*args, **kwargs) 2025-09-07T08:12:20.0723844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T08:12:20.0724247Z self.key(current_states) 2025-09-07T08:12:20.0724354Z 2025-09-07T08:12:20.0724451Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0724798Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0725113Z return mod(**inputs) 2025-09-07T08:12:20.0725492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0725894Z outputs = self.roberta( 2025-09-07T08:12:20.0726268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0726673Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0727066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0727466Z layer_outputs = layer_module( 2025-09-07T08:12:20.0727800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0728147Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0728550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0728964Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0729333Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0729690Z return func(*args, **kwargs) 2025-09-07T08:12:20.0730074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0730474Z self_outputs = self.self( 2025-09-07T08:12:20.0730822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0731182Z return func(*args, **kwargs) 2025-09-07T08:12:20.0731560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T08:12:20.0732000Z self.value(current_states) 2025-09-07T08:12:20.0732148Z 2025-09-07T08:12:20.0732224Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0732449Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0732795Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0733105Z return mod(**inputs) 2025-09-07T08:12:20.0733484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0733893Z outputs = self.roberta( 2025-09-07T08:12:20.0734270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0734666Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0735062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0735468Z layer_outputs = layer_module( 2025-09-07T08:12:20.0735842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0736200Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0736630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0737053Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0737426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0737846Z return func(*args, **kwargs) 2025-09-07T08:12:20.0738241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0738649Z self_outputs = self.self( 2025-09-07T08:12:20.0739007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0739375Z return func(*args, **kwargs) 2025-09-07T08:12:20.0739759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:12:20.0740215Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:12:20.0740403Z 2025-09-07T08:12:20.0740503Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0740858Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0741175Z return mod(**inputs) 2025-09-07T08:12:20.0741555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0741957Z outputs = self.roberta( 2025-09-07T08:12:20.0742335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0742743Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0743134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0743539Z layer_outputs = layer_module( 2025-09-07T08:12:20.0743867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0744219Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0744622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0745033Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0745395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0745833Z return func(*args, **kwargs) 2025-09-07T08:12:20.0746256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T08:12:20.0746712Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:12:20.0747166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T08:12:20.0747582Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0747722Z 2025-09-07T08:12:20.0747816Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0748159Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0748477Z return mod(**inputs) 2025-09-07T08:12:20.0748854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0749253Z outputs = self.roberta( 2025-09-07T08:12:20.0749680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0750088Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0750520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0750922Z layer_outputs = layer_module( 2025-09-07T08:12:20.0751268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0751617Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0752019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0752434Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0752826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0753230Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0753666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0754152Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0754602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T08:12:20.0755014Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0755154Z 2025-09-07T08:12:20.0755249Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0755594Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0755912Z return mod(**inputs) 2025-09-07T08:12:20.0756287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0756689Z outputs = self.roberta( 2025-09-07T08:12:20.0757067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0757473Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0757866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0758264Z layer_outputs = layer_module( 2025-09-07T08:12:20.0758598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0758952Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0759352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0759811Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0760248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0760645Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0761077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0761557Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0762001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T08:12:20.0762439Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:12:20.0762817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:12:20.0763155Z return self.act(input) 2025-09-07T08:12:20.0763261Z 2025-09-07T08:12:20.0763364Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0763751Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0764066Z return mod(**inputs) 2025-09-07T08:12:20.0764486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0764903Z outputs = self.roberta( 2025-09-07T08:12:20.0765285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0765687Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0766079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0766486Z layer_outputs = layer_module( 2025-09-07T08:12:20.0766821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0767180Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0767573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0767985Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0768379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0768771Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0769205Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T08:12:20.0769695Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:12:20.0770166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T08:12:20.0770591Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0770731Z 2025-09-07T08:12:20.0770841Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0771198Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0771524Z return mod(**inputs) 2025-09-07T08:12:20.0771912Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0772321Z outputs = self.roberta( 2025-09-07T08:12:20.0772708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0773114Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0773515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0773970Z layer_outputs = layer_module( 2025-09-07T08:12:20.0774346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0774713Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0775117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0775532Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0775916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0776293Z return func(*args, **kwargs) 2025-09-07T08:12:20.0776683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0777077Z self_outputs = self.self( 2025-09-07T08:12:20.0777430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0777885Z return func(*args, **kwargs) 2025-09-07T08:12:20.0778274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T08:12:20.0778844Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:12:20.0779119Z 2025-09-07T08:12:20.0779219Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0779577Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0779898Z return mod(**inputs) 2025-09-07T08:12:20.0780280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0780683Z outputs = self.roberta( 2025-09-07T08:12:20.0781072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0781479Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0781880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0782288Z layer_outputs = layer_module( 2025-09-07T08:12:20.0782621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0782974Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0783396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0783823Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0784202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0784582Z return func(*args, **kwargs) 2025-09-07T08:12:20.0784983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0785391Z self_outputs = self.self( 2025-09-07T08:12:20.0785749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0786109Z return func(*args, **kwargs) 2025-09-07T08:12:20.0786491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T08:12:20.0786902Z self.key(current_states) 2025-09-07T08:12:20.0787009Z 2025-09-07T08:12:20.0787117Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0787476Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0787792Z return mod(**inputs) 2025-09-07T08:12:20.0788229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0788677Z outputs = self.roberta( 2025-09-07T08:12:20.0789067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0789468Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0789872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0790278Z layer_outputs = layer_module( 2025-09-07T08:12:20.0790620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0790985Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0791386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0792054Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0792434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0792807Z return func(*args, **kwargs) 2025-09-07T08:12:20.0793228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0793630Z self_outputs = self.self( 2025-09-07T08:12:20.0793982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0794348Z return func(*args, **kwargs) 2025-09-07T08:12:20.0794728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T08:12:20.0795241Z self.value(current_states) 2025-09-07T08:12:20.0795361Z 2025-09-07T08:12:20.0795437Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0795674Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0796026Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0796344Z return mod(**inputs) 2025-09-07T08:12:20.0796726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0797131Z outputs = self.roberta( 2025-09-07T08:12:20.0797510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0797914Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0798309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0798703Z layer_outputs = layer_module( 2025-09-07T08:12:20.0799046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0799408Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0799814Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0800222Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0800595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0800960Z return func(*args, **kwargs) 2025-09-07T08:12:20.0801345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0801746Z self_outputs = self.self( 2025-09-07T08:12:20.0802084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0802517Z return func(*args, **kwargs) 2025-09-07T08:12:20.0802937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:12:20.0803537Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:12:20.0803721Z 2025-09-07T08:12:20.0803828Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0804172Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0804484Z return mod(**inputs) 2025-09-07T08:12:20.0804866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0805272Z outputs = self.roberta( 2025-09-07T08:12:20.0805646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0806051Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0806528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0806944Z layer_outputs = layer_module( 2025-09-07T08:12:20.0807355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0807717Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0808131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0808545Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0808926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0809297Z return func(*args, **kwargs) 2025-09-07T08:12:20.0809683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T08:12:20.0810148Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:12:20.0810607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T08:12:20.0811029Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0811166Z 2025-09-07T08:12:20.0811268Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0811611Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0811928Z return mod(**inputs) 2025-09-07T08:12:20.0812311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0812718Z outputs = self.roberta( 2025-09-07T08:12:20.0813091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0813502Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0813902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0814314Z layer_outputs = layer_module( 2025-09-07T08:12:20.0814650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0815002Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0815406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0815828Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0816229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0816702Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0817179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0817721Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0818174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T08:12:20.0818591Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0818722Z 2025-09-07T08:12:20.0818817Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0819164Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0819484Z return mod(**inputs) 2025-09-07T08:12:20.0819864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0820274Z outputs = self.roberta( 2025-09-07T08:12:20.0820683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0821094Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0821519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0821936Z layer_outputs = layer_module( 2025-09-07T08:12:20.0822285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0822637Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0823040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0823449Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0823842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0824228Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0824665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0825137Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0825572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T08:12:20.0826006Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:12:20.0826369Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:12:20.0826699Z return self.act(input) 2025-09-07T08:12:20.0826803Z 2025-09-07T08:12:20.0826895Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0827240Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0827549Z return mod(**inputs) 2025-09-07T08:12:20.0827918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0828319Z outputs = self.roberta( 2025-09-07T08:12:20.0828691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0829091Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0829480Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0829874Z layer_outputs = layer_module( 2025-09-07T08:12:20.0830206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0830603Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0831039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0831449Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0831843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0840082Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0840584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T08:12:20.0841114Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:12:20.0841582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T08:12:20.0842013Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0842161Z 2025-09-07T08:12:20.0842359Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0842726Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0843047Z return mod(**inputs) 2025-09-07T08:12:20.0843477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0843893Z outputs = self.roberta( 2025-09-07T08:12:20.0844287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0844703Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0845096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0845494Z layer_outputs = layer_module( 2025-09-07T08:12:20.0845844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0846212Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0846622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0847035Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0847409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0847775Z return func(*args, **kwargs) 2025-09-07T08:12:20.0848164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0848561Z self_outputs = self.self( 2025-09-07T08:12:20.0848905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0849261Z return func(*args, **kwargs) 2025-09-07T08:12:20.0849665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T08:12:20.0850217Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:12:20.0850481Z 2025-09-07T08:12:20.0850594Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0850939Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0851262Z return mod(**inputs) 2025-09-07T08:12:20.0851637Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0852033Z outputs = self.roberta( 2025-09-07T08:12:20.0852405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0852908Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0853303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0853700Z layer_outputs = layer_module( 2025-09-07T08:12:20.0854034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0854385Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0854779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0855182Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0855548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0855911Z return func(*args, **kwargs) 2025-09-07T08:12:20.0856327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0856734Z self_outputs = self.self( 2025-09-07T08:12:20.0857082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0857474Z return func(*args, **kwargs) 2025-09-07T08:12:20.0857919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T08:12:20.0858318Z self.key(current_states) 2025-09-07T08:12:20.0858430Z 2025-09-07T08:12:20.0858524Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0858873Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0859185Z return mod(**inputs) 2025-09-07T08:12:20.0859555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0859965Z outputs = self.roberta( 2025-09-07T08:12:20.0860338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0860758Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0861168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0861569Z layer_outputs = layer_module( 2025-09-07T08:12:20.0861906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0862263Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0862668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0863079Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0863449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0863814Z return func(*args, **kwargs) 2025-09-07T08:12:20.0864207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0864606Z self_outputs = self.self( 2025-09-07T08:12:20.0864951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0865309Z return func(*args, **kwargs) 2025-09-07T08:12:20.0865685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T08:12:20.0866082Z self.value(current_states) 2025-09-07T08:12:20.0866191Z 2025-09-07T08:12:20.0866273Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0866541Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0866936Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0867253Z return mod(**inputs) 2025-09-07T08:12:20.0867628Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0868021Z outputs = self.roberta( 2025-09-07T08:12:20.0868391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0868791Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0869179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0869578Z layer_outputs = layer_module( 2025-09-07T08:12:20.0869909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0870295Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0870700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0871115Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0871539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0871910Z return func(*args, **kwargs) 2025-09-07T08:12:20.0872300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0872697Z self_outputs = self.self( 2025-09-07T08:12:20.0873044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0873401Z return func(*args, **kwargs) 2025-09-07T08:12:20.0873781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:12:20.0874240Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:12:20.0874420Z 2025-09-07T08:12:20.0874518Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0874864Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0875173Z return mod(**inputs) 2025-09-07T08:12:20.0875550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0875957Z outputs = self.roberta( 2025-09-07T08:12:20.0876335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0876740Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0877130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0877532Z layer_outputs = layer_module( 2025-09-07T08:12:20.0877870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0878227Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0878629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0879036Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0879401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0879767Z return func(*args, **kwargs) 2025-09-07T08:12:20.0880148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T08:12:20.0880644Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:12:20.0881129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T08:12:20.0881536Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0881669Z 2025-09-07T08:12:20.0881766Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0882108Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0882412Z return mod(**inputs) 2025-09-07T08:12:20.0882779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0883179Z outputs = self.roberta( 2025-09-07T08:12:20.0883547Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0883945Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0884412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0884813Z layer_outputs = layer_module( 2025-09-07T08:12:20.0885177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0885526Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0885916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0886322Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0886712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0887096Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0887525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0888004Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0888445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T08:12:20.0888848Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0888976Z 2025-09-07T08:12:20.0889072Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0889409Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0889712Z return mod(**inputs) 2025-09-07T08:12:20.0890080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0890476Z outputs = self.roberta( 2025-09-07T08:12:20.0890848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0891241Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0891626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0892017Z layer_outputs = layer_module( 2025-09-07T08:12:20.0892344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0892696Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0893086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0893493Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0893885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0894343Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0894771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0895245Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0895686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T08:12:20.0896124Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:12:20.0896501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:12:20.0896835Z return self.act(input) 2025-09-07T08:12:20.0896936Z 2025-09-07T08:12:20.0897033Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0897378Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0897730Z return mod(**inputs) 2025-09-07T08:12:20.0898149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0898557Z outputs = self.roberta( 2025-09-07T08:12:20.0898966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0899379Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0899761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0900152Z layer_outputs = layer_module( 2025-09-07T08:12:20.0900476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0900821Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0901214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0901291Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0901549Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0901617Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0901924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T08:12:20.0902047Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:12:20.0902317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T08:12:20.0902386Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0902390Z 2025-09-07T08:12:20.0902486Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0902679Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0902738Z return mod(**inputs) 2025-09-07T08:12:20.0903015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0903237Z outputs = self.roberta( 2025-09-07T08:12:20.0903529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0903594Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0903867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0903927Z layer_outputs = layer_module( 2025-09-07T08:12:20.0904144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0904332Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0904654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0904731Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0904978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0905042Z return func(*args, **kwargs) 2025-09-07T08:12:20.0905313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0905375Z self_outputs = self.self( 2025-09-07T08:12:20.0905611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0905668Z return func(*args, **kwargs) 2025-09-07T08:12:20.0905987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T08:12:20.0906200Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:12:20.0906204Z 2025-09-07T08:12:20.0906342Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0906545Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0906606Z return mod(**inputs) 2025-09-07T08:12:20.0906895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0906958Z outputs = self.roberta( 2025-09-07T08:12:20.0907233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0907297Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0907569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0907635Z layer_outputs = layer_module( 2025-09-07T08:12:20.0907854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0907925Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0908197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0908277Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0908519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0908581Z return func(*args, **kwargs) 2025-09-07T08:12:20.0908855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0908920Z self_outputs = self.self( 2025-09-07T08:12:20.0909160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0909219Z return func(*args, **kwargs) 2025-09-07T08:12:20.0909492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T08:12:20.0909556Z self.key(current_states) 2025-09-07T08:12:20.0909559Z 2025-09-07T08:12:20.0909654Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0909849Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0909907Z return mod(**inputs) 2025-09-07T08:12:20.0910185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0910286Z outputs = self.roberta( 2025-09-07T08:12:20.0910593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0910658Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0910931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0910994Z layer_outputs = layer_module( 2025-09-07T08:12:20.0911210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0911277Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0911552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0911624Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0911864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0911980Z return func(*args, **kwargs) 2025-09-07T08:12:20.0912254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0912348Z self_outputs = self.self( 2025-09-07T08:12:20.0912591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0912657Z return func(*args, **kwargs) 2025-09-07T08:12:20.0912927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T08:12:20.0912992Z self.value(current_states) 2025-09-07T08:12:20.0912995Z 2025-09-07T08:12:20.0913068Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0913160Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0913354Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0913412Z return mod(**inputs) 2025-09-07T08:12:20.0913687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0913746Z outputs = self.roberta( 2025-09-07T08:12:20.0914016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0914083Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0914356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0914426Z layer_outputs = layer_module( 2025-09-07T08:12:20.0914641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0914711Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0914990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0915062Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0915308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0915368Z return func(*args, **kwargs) 2025-09-07T08:12:20.0915643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0915703Z self_outputs = self.self( 2025-09-07T08:12:20.0915940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0916005Z return func(*args, **kwargs) 2025-09-07T08:12:20.0916278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:12:20.0916481Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:12:20.0916484Z 2025-09-07T08:12:20.0916573Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0916759Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0916816Z return mod(**inputs) 2025-09-07T08:12:20.0917091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0917150Z outputs = self.roberta( 2025-09-07T08:12:20.0917420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0917488Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0917758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0917825Z layer_outputs = layer_module( 2025-09-07T08:12:20.0918074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0918142Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0918452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0918521Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0918754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0918814Z return func(*args, **kwargs) 2025-09-07T08:12:20.0919082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T08:12:20.0919203Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:12:20.0919474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T08:12:20.0919554Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0919558Z 2025-09-07T08:12:20.0919649Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0919833Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0919894Z return mod(**inputs) 2025-09-07T08:12:20.0920168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0920229Z outputs = self.roberta( 2025-09-07T08:12:20.0920500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0920562Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0920837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0920898Z layer_outputs = layer_module( 2025-09-07T08:12:20.0921116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0921181Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0921449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0921527Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0921788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0921859Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0922166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0922344Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0922614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T08:12:20.0922686Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0922690Z 2025-09-07T08:12:20.0922780Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0922966Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0923028Z return mod(**inputs) 2025-09-07T08:12:20.0923302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0923359Z outputs = self.roberta( 2025-09-07T08:12:20.0923638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0923707Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0924018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0924081Z layer_outputs = layer_module( 2025-09-07T08:12:20.0924331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0924399Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0924671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0924748Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0925005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0925074Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0925381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0925488Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0925765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T08:12:20.0925867Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:12:20.0926079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:12:20.0926141Z return self.act(input) 2025-09-07T08:12:20.0926144Z 2025-09-07T08:12:20.0926239Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0926423Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0926477Z return mod(**inputs) 2025-09-07T08:12:20.0926761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0926822Z outputs = self.roberta( 2025-09-07T08:12:20.0927098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0927161Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0927438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0927506Z layer_outputs = layer_module( 2025-09-07T08:12:20.0927718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0927789Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0928057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0928195Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0928454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0928519Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0928825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T08:12:20.0928948Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:12:20.0929220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T08:12:20.0929293Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0929296Z 2025-09-07T08:12:20.0929389Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0929578Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0929664Z return mod(**inputs) 2025-09-07T08:12:20.0929958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0930018Z outputs = self.roberta( 2025-09-07T08:12:20.0930321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0930385Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0930657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0930726Z layer_outputs = layer_module( 2025-09-07T08:12:20.0930939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0931005Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0931276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0931347Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0931589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0931649Z return func(*args, **kwargs) 2025-09-07T08:12:20.0931920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0931980Z self_outputs = self.self( 2025-09-07T08:12:20.0932219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0932280Z return func(*args, **kwargs) 2025-09-07T08:12:20.0932548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T08:12:20.0932757Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:12:20.0932760Z 2025-09-07T08:12:20.0932851Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0933043Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0933100Z return mod(**inputs) 2025-09-07T08:12:20.0933375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0933439Z outputs = self.roberta( 2025-09-07T08:12:20.0933708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0933772Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0934042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0934401Z layer_outputs = layer_module( 2025-09-07T08:12:20.0934617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0934682Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0934955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0935024Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0935261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0935317Z return func(*args, **kwargs) 2025-09-07T08:12:20.0935584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0935649Z self_outputs = self.self( 2025-09-07T08:12:20.0935925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0935992Z return func(*args, **kwargs) 2025-09-07T08:12:20.0936298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T08:12:20.0936359Z self.key(current_states) 2025-09-07T08:12:20.0936364Z 2025-09-07T08:12:20.0936460Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0936652Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0936712Z return mod(**inputs) 2025-09-07T08:12:20.0936990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0937057Z outputs = self.roberta( 2025-09-07T08:12:20.0937331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0937400Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0937723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0937788Z layer_outputs = layer_module( 2025-09-07T08:12:20.0938011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0938079Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0938350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0938423Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0938658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0938723Z return func(*args, **kwargs) 2025-09-07T08:12:20.0938996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0939058Z self_outputs = self.self( 2025-09-07T08:12:20.0939292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0939350Z return func(*args, **kwargs) 2025-09-07T08:12:20.0939621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T08:12:20.0939683Z self.value(current_states) 2025-09-07T08:12:20.0939686Z 2025-09-07T08:12:20.0939759Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0939848Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0940033Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0940132Z return mod(**inputs) 2025-09-07T08:12:20.0940442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0940504Z outputs = self.roberta( 2025-09-07T08:12:20.0940775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0940837Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0941110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0941172Z layer_outputs = layer_module( 2025-09-07T08:12:20.0941388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0941454Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0941724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0941828Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0942070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0942164Z return func(*args, **kwargs) 2025-09-07T08:12:20.0942439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0942506Z self_outputs = self.self( 2025-09-07T08:12:20.0942746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0942808Z return func(*args, **kwargs) 2025-09-07T08:12:20.0943086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:12:20.0943209Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:12:20.0943214Z 2025-09-07T08:12:20.0943316Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0943504Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0943568Z return mod(**inputs) 2025-09-07T08:12:20.0943846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0943903Z outputs = self.roberta( 2025-09-07T08:12:20.0944179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0944243Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0944513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0944575Z layer_outputs = layer_module( 2025-09-07T08:12:20.0944794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0944868Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0945136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0945209Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0945445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0945503Z return func(*args, **kwargs) 2025-09-07T08:12:20.0945770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T08:12:20.0945886Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:12:20.0946157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T08:12:20.0946293Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0946296Z 2025-09-07T08:12:20.0946387Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0946572Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0946627Z return mod(**inputs) 2025-09-07T08:12:20.0946904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0946961Z outputs = self.roberta( 2025-09-07T08:12:20.0947233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0947295Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0947563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0947629Z layer_outputs = layer_module( 2025-09-07T08:12:20.0947874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0947944Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0948247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0948321Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0948586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0948655Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0948966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0949079Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0949352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T08:12:20.0949420Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0949424Z 2025-09-07T08:12:20.0949515Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0949698Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0949753Z return mod(**inputs) 2025-09-07T08:12:20.0950029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0950087Z outputs = self.roberta( 2025-09-07T08:12:20.0950356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0950416Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0950688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0950752Z layer_outputs = layer_module( 2025-09-07T08:12:20.0950965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0951032Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0951299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0951370Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0951631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0951696Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0952001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0952164Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0952437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T08:12:20.0952538Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:12:20.0952746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:12:20.0952807Z return self.act(input) 2025-09-07T08:12:20.0952811Z 2025-09-07T08:12:20.0952901Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0953086Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0953143Z return mod(**inputs) 2025-09-07T08:12:20.0953421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0953486Z outputs = self.roberta( 2025-09-07T08:12:20.0953790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0953860Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0954166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0954231Z layer_outputs = layer_module( 2025-09-07T08:12:20.0954445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0954513Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0954786Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0954854Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0955119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0955183Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0955490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T08:12:20.0955616Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:12:20.0955888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T08:12:20.0955959Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0955963Z 2025-09-07T08:12:20.0956049Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0956235Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0956290Z return mod(**inputs) 2025-09-07T08:12:20.0956564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0956627Z outputs = self.roberta( 2025-09-07T08:12:20.0956899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0956962Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0957229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0957287Z layer_outputs = layer_module( 2025-09-07T08:12:20.0957500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0957564Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0957834Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0957977Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0958221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0958282Z return func(*args, **kwargs) 2025-09-07T08:12:20.0958556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0958631Z self_outputs = self.self( 2025-09-07T08:12:20.0958871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0958936Z return func(*args, **kwargs) 2025-09-07T08:12:20.0959206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T08:12:20.0959409Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:12:20.0959422Z 2025-09-07T08:12:20.0959546Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0959733Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0959836Z return mod(**inputs) 2025-09-07T08:12:20.0960115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0960185Z outputs = self.roberta( 2025-09-07T08:12:20.0960457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0960523Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0960801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0960864Z layer_outputs = layer_module( 2025-09-07T08:12:20.0961093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0961162Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0961437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0961522Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0961761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0961832Z return func(*args, **kwargs) 2025-09-07T08:12:20.0962107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0962180Z self_outputs = self.self( 2025-09-07T08:12:20.0962419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0962482Z return func(*args, **kwargs) 2025-09-07T08:12:20.0962760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T08:12:20.0962820Z self.key(current_states) 2025-09-07T08:12:20.0962825Z 2025-09-07T08:12:20.0962923Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0963109Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0963167Z return mod(**inputs) 2025-09-07T08:12:20.0963447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0963507Z outputs = self.roberta( 2025-09-07T08:12:20.0963783Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0963880Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0964184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0964254Z layer_outputs = layer_module( 2025-09-07T08:12:20.0964470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0964543Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0964812Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0964892Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0965128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0965191Z return func(*args, **kwargs) 2025-09-07T08:12:20.0965472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0965567Z self_outputs = self.self( 2025-09-07T08:12:20.0965816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0965905Z return func(*args, **kwargs) 2025-09-07T08:12:20.0966179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T08:12:20.0966250Z self.value(current_states) 2025-09-07T08:12:20.0966253Z 2025-09-07T08:12:20.0966327Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0966425Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0966613Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0966670Z return mod(**inputs) 2025-09-07T08:12:20.0966951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0967014Z outputs = self.roberta( 2025-09-07T08:12:20.0967295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0967359Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0967635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0967697Z layer_outputs = layer_module( 2025-09-07T08:12:20.0967909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0967981Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0968252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0968331Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0968569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0968630Z return func(*args, **kwargs) 2025-09-07T08:12:20.0968909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0968969Z self_outputs = self.self( 2025-09-07T08:12:20.0969212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0969274Z return func(*args, **kwargs) 2025-09-07T08:12:20.0969550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:12:20.0969675Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:12:20.0969709Z 2025-09-07T08:12:20.0969799Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0970021Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0970079Z return mod(**inputs) 2025-09-07T08:12:20.0970364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0970423Z outputs = self.roberta( 2025-09-07T08:12:20.0970694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0970764Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0971035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0971105Z layer_outputs = layer_module( 2025-09-07T08:12:20.0971317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0971425Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0971699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0971803Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0972056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0972119Z return func(*args, **kwargs) 2025-09-07T08:12:20.0972402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T08:12:20.0972522Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:12:20.0972796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T08:12:20.0972884Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0972889Z 2025-09-07T08:12:20.0972984Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0973182Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0973239Z return mod(**inputs) 2025-09-07T08:12:20.0973533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0973594Z outputs = self.roberta( 2025-09-07T08:12:20.0973868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0973942Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0974217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0974290Z layer_outputs = layer_module( 2025-09-07T08:12:20.0974511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0974586Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0974866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0974942Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0975211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0975281Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0975593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0975717Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0976002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T08:12:20.0976147Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0976150Z 2025-09-07T08:12:20.0976250Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0976455Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0976515Z return mod(**inputs) 2025-09-07T08:12:20.0976798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0976866Z outputs = self.roberta( 2025-09-07T08:12:20.0977137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0977208Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0977479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0977639Z layer_outputs = layer_module( 2025-09-07T08:12:20.0977872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0977947Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0978277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0978356Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0978624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0978695Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0979004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.0979128Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.0979406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T08:12:20.0979517Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:12:20.0979730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:12:20.0979794Z return self.act(input) 2025-09-07T08:12:20.0979802Z 2025-09-07T08:12:20.0979903Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0980096Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0980161Z return mod(**inputs) 2025-09-07T08:12:20.0980440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0980509Z outputs = self.roberta( 2025-09-07T08:12:20.0980785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0980853Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0981132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0981197Z layer_outputs = layer_module( 2025-09-07T08:12:20.0981421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0981493Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0981766Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.0981848Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.0982107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.0982240Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.0982546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T08:12:20.0982683Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:12:20.0982955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T08:12:20.0983027Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0983031Z 2025-09-07T08:12:20.0983132Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0983321Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0983385Z return mod(**inputs) 2025-09-07T08:12:20.0983661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0983755Z outputs = self.roberta( 2025-09-07T08:12:20.0984039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0984132Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0984413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0984476Z layer_outputs = layer_module( 2025-09-07T08:12:20.0984699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0984769Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0985050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0985134Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0985381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0985451Z return func(*args, **kwargs) 2025-09-07T08:12:20.0985723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0985788Z self_outputs = self.self( 2025-09-07T08:12:20.0986034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0986094Z return func(*args, **kwargs) 2025-09-07T08:12:20.0986373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T08:12:20.0986575Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:12:20.0986580Z 2025-09-07T08:12:20.0986670Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0986863Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0986918Z return mod(**inputs) 2025-09-07T08:12:20.0987198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0987258Z outputs = self.roberta( 2025-09-07T08:12:20.0987532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0987594Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0987864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0987927Z layer_outputs = layer_module( 2025-09-07T08:12:20.0988140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0988316Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0988589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0988665Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0988902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0988964Z return func(*args, **kwargs) 2025-09-07T08:12:20.0989237Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0989298Z self_outputs = self.self( 2025-09-07T08:12:20.0989534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0989593Z return func(*args, **kwargs) 2025-09-07T08:12:20.0989896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T08:12:20.0989964Z self.key(current_states) 2025-09-07T08:12:20.0989967Z 2025-09-07T08:12:20.0990061Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0990286Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0990342Z return mod(**inputs) 2025-09-07T08:12:20.0990618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0990681Z outputs = self.roberta( 2025-09-07T08:12:20.0990950Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0991014Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0991286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0991356Z layer_outputs = layer_module( 2025-09-07T08:12:20.0991570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0991636Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0991909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0991981Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0992219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0992278Z return func(*args, **kwargs) 2025-09-07T08:12:20.0992546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0992616Z self_outputs = self.self( 2025-09-07T08:12:20.0992852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0992912Z return func(*args, **kwargs) 2025-09-07T08:12:20.0993182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T08:12:20.0993250Z self.value(current_states) 2025-09-07T08:12:20.0993253Z 2025-09-07T08:12:20.0993325Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.0993413Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0993599Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0993654Z return mod(**inputs) 2025-09-07T08:12:20.0993932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0994024Z outputs = self.roberta( 2025-09-07T08:12:20.0994325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0994396Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0994669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0994732Z layer_outputs = layer_module( 2025-09-07T08:12:20.0994945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0995011Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0995285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0995353Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0995623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0995685Z return func(*args, **kwargs) 2025-09-07T08:12:20.0995960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.0996052Z self_outputs = self.self( 2025-09-07T08:12:20.0996289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0996350Z return func(*args, **kwargs) 2025-09-07T08:12:20.0996621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:12:20.0996750Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:12:20.0996753Z 2025-09-07T08:12:20.0996842Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.0997026Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.0997089Z return mod(**inputs) 2025-09-07T08:12:20.0997363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.0997426Z outputs = self.roberta( 2025-09-07T08:12:20.0997693Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.0997758Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.0998027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.0998085Z layer_outputs = layer_module( 2025-09-07T08:12:20.0998298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.0998364Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.0998639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.0998708Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.0998942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.0999003Z return func(*args, **kwargs) 2025-09-07T08:12:20.0999271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T08:12:20.0999394Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:12:20.0999665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T08:12:20.0999742Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.0999778Z 2025-09-07T08:12:20.0999868Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1000083Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1000140Z return mod(**inputs) 2025-09-07T08:12:20.1000418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1000480Z outputs = self.roberta( 2025-09-07T08:12:20.1000751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1000812Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1001086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1001145Z layer_outputs = layer_module( 2025-09-07T08:12:20.1001359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1001475Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1001747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.1001849Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.1002113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.1002186Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.1002491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.1002608Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.1002878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T08:12:20.1002953Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.1002958Z 2025-09-07T08:12:20.1003204Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1003391Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1003452Z return mod(**inputs) 2025-09-07T08:12:20.1003730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1003789Z outputs = self.roberta( 2025-09-07T08:12:20.1004063Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1004125Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1004408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1004470Z layer_outputs = layer_module( 2025-09-07T08:12:20.1004686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1004751Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1005020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.1005091Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.1005348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.1005415Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.1005715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.1005817Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.1006152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T08:12:20.1006297Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:12:20.1006511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:12:20.1006571Z return self.act(input) 2025-09-07T08:12:20.1006575Z 2025-09-07T08:12:20.1006674Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1006865Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1006923Z return mod(**inputs) 2025-09-07T08:12:20.1007212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1007274Z outputs = self.roberta( 2025-09-07T08:12:20.1007548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1007659Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1007933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1008038Z layer_outputs = layer_module( 2025-09-07T08:12:20.1008259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1008330Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1008599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.1008674Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.1008936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.1009004Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.1009314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T08:12:20.1009440Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:12:20.1009715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T08:12:20.1009787Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.1009790Z 2025-09-07T08:12:20.1009884Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1010067Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1010123Z return mod(**inputs) 2025-09-07T08:12:20.1010398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1010457Z outputs = self.roberta( 2025-09-07T08:12:20.1010733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1010793Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1011062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1011126Z layer_outputs = layer_module( 2025-09-07T08:12:20.1011341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1011410Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1011681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.1011756Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.1012006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1012136Z return func(*args, **kwargs) 2025-09-07T08:12:20.1012417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.1012483Z self_outputs = self.self( 2025-09-07T08:12:20.1012728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1012789Z return func(*args, **kwargs) 2025-09-07T08:12:20.1013063Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T08:12:20.1013275Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:12:20.1013279Z 2025-09-07T08:12:20.1013370Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1013595Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1013656Z return mod(**inputs) 2025-09-07T08:12:20.1013941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1014039Z outputs = self.roberta( 2025-09-07T08:12:20.1014317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1014389Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1014662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1014732Z layer_outputs = layer_module( 2025-09-07T08:12:20.1014949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1015018Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1015301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.1015370Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.1015620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1015682Z return func(*args, **kwargs) 2025-09-07T08:12:20.1015958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.1016025Z self_outputs = self.self( 2025-09-07T08:12:20.1016260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1016329Z return func(*args, **kwargs) 2025-09-07T08:12:20.1016598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T08:12:20.1016666Z self.key(current_states) 2025-09-07T08:12:20.1016674Z 2025-09-07T08:12:20.1016767Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1016959Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1017023Z return mod(**inputs) 2025-09-07T08:12:20.1017315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1017381Z outputs = self.roberta( 2025-09-07T08:12:20.1017705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1017770Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1018050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1018174Z layer_outputs = layer_module( 2025-09-07T08:12:20.1018408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1018480Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1018760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.1018840Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.1019079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1019148Z return func(*args, **kwargs) 2025-09-07T08:12:20.1019418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.1019487Z self_outputs = self.self( 2025-09-07T08:12:20.1019756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1019818Z return func(*args, **kwargs) 2025-09-07T08:12:20.1020096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T08:12:20.1020189Z self.value(current_states) 2025-09-07T08:12:20.1020192Z 2025-09-07T08:12:20.1020273Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.1020374Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1020572Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1020636Z return mod(**inputs) 2025-09-07T08:12:20.1020911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1020976Z outputs = self.roberta( 2025-09-07T08:12:20.1021249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1021314Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1021587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1021651Z layer_outputs = layer_module( 2025-09-07T08:12:20.1021872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1021939Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1022214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.1022286Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.1022521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1022586Z return func(*args, **kwargs) 2025-09-07T08:12:20.1022861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.1022927Z self_outputs = self.self( 2025-09-07T08:12:20.1023166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1023227Z return func(*args, **kwargs) 2025-09-07T08:12:20.1023502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:12:20.1023626Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:12:20.1023630Z 2025-09-07T08:12:20.1023724Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1023910Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1024004Z return mod(**inputs) 2025-09-07T08:12:20.1024331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1024392Z outputs = self.roberta( 2025-09-07T08:12:20.1024675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1024737Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1025010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1025074Z layer_outputs = layer_module( 2025-09-07T08:12:20.1025289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1025362Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1025630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.1025758Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.1026001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1026092Z return func(*args, **kwargs) 2025-09-07T08:12:20.1026372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T08:12:20.1026491Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:12:20.1026767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T08:12:20.1026839Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.1026842Z 2025-09-07T08:12:20.1026940Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1027130Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1027190Z return mod(**inputs) 2025-09-07T08:12:20.1027467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1027530Z outputs = self.roberta( 2025-09-07T08:12:20.1027807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1027870Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1028137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1028201Z layer_outputs = layer_module( 2025-09-07T08:12:20.1028413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1028486Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1028756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.1028832Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.1029098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.1029165Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.1029473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.1029583Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.1029859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T08:12:20.1029931Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.1029966Z 2025-09-07T08:12:20.1030058Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1030274Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1030333Z return mod(**inputs) 2025-09-07T08:12:20.1030620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1030686Z outputs = self.roberta( 2025-09-07T08:12:20.1030961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1031026Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1031293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1031361Z layer_outputs = layer_module( 2025-09-07T08:12:20.1031573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1031676Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1031950Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.1032054Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.1032324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.1032391Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.1032707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.1032812Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.1033104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T08:12:20.1033209Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:12:20.1033418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:12:20.1033485Z return self.act(input) 2025-09-07T08:12:20.1033489Z 2025-09-07T08:12:20.1033579Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1033769Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1033824Z return mod(**inputs) 2025-09-07T08:12:20.1034101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1034170Z outputs = self.roberta( 2025-09-07T08:12:20.1034443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1034516Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1034790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1034860Z layer_outputs = layer_module( 2025-09-07T08:12:20.1035074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1035142Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1035417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.1035488Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.1035750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.1035814Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.1036118Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T08:12:20.1036310Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:12:20.1036581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T08:12:20.1036658Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.1036661Z 2025-09-07T08:12:20.1036751Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1036940Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1036995Z return mod(**inputs) 2025-09-07T08:12:20.1037270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1037337Z outputs = self.roberta( 2025-09-07T08:12:20.1037640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1037710Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1037978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1038071Z layer_outputs = layer_module( 2025-09-07T08:12:20.1038291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1038360Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1038638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.1038709Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.1038952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1039015Z return func(*args, **kwargs) 2025-09-07T08:12:20.1039289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.1039355Z self_outputs = self.self( 2025-09-07T08:12:20.1039593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1039653Z return func(*args, **kwargs) 2025-09-07T08:12:20.1039923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 325, in forward 2025-09-07T08:12:20.1040126Z self.query(hidden_states).view(bsz, -1, self.num_attention_heads, self.attention_head_size).transpose(1, 2) 2025-09-07T08:12:20.1040135Z 2025-09-07T08:12:20.1040228Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1040414Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1040478Z return mod(**inputs) 2025-09-07T08:12:20.1040755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1040820Z outputs = self.roberta( 2025-09-07T08:12:20.1041092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1041156Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1041429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1041492Z layer_outputs = layer_module( 2025-09-07T08:12:20.1041709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1041776Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1042086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.1042192Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.1042429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1042498Z return func(*args, **kwargs) 2025-09-07T08:12:20.1042771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.1042834Z self_outputs = self.self( 2025-09-07T08:12:20.1043071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1043131Z return func(*args, **kwargs) 2025-09-07T08:12:20.1043408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 353, in forward 2025-09-07T08:12:20.1043472Z self.key(current_states) 2025-09-07T08:12:20.1043477Z 2025-09-07T08:12:20.1043602Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1043785Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1043844Z return mod(**inputs) 2025-09-07T08:12:20.1044158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1044217Z outputs = self.roberta( 2025-09-07T08:12:20.1044492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1044556Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1044827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1044896Z layer_outputs = layer_module( 2025-09-07T08:12:20.1045115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1045189Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1045461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.1045535Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.1045769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1045830Z return func(*args, **kwargs) 2025-09-07T08:12:20.1046104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.1046164Z self_outputs = self.self( 2025-09-07T08:12:20.1046404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1046465Z return func(*args, **kwargs) 2025-09-07T08:12:20.1046737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 358, in forward 2025-09-07T08:12:20.1046802Z self.value(current_states) 2025-09-07T08:12:20.1046805Z 2025-09-07T08:12:20.1046879Z cudagraph partition due to non gpu ops 2025-09-07T08:12:20.1046977Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1047157Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1047215Z return mod(**inputs) 2025-09-07T08:12:20.1047496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1047557Z outputs = self.roberta( 2025-09-07T08:12:20.1047837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1047975Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1048254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1048314Z layer_outputs = layer_module( 2025-09-07T08:12:20.1048527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1048600Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1048869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.1048947Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.1049183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1049244Z return func(*args, **kwargs) 2025-09-07T08:12:20.1049555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 467, in forward 2025-09-07T08:12:20.1049617Z self_outputs = self.self( 2025-09-07T08:12:20.1049890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1049949Z return func(*args, **kwargs) 2025-09-07T08:12:20.1050227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 389, in forward 2025-09-07T08:12:20.1050350Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:12:20.1050353Z 2025-09-07T08:12:20.1050443Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1050634Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1050691Z return mod(**inputs) 2025-09-07T08:12:20.1050976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1051036Z outputs = self.roberta( 2025-09-07T08:12:20.1051307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1051377Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1051648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1051716Z layer_outputs = layer_module( 2025-09-07T08:12:20.1051933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1052005Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1052272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 540, in forward 2025-09-07T08:12:20.1052345Z self_attention_outputs = self.attention( 2025-09-07T08:12:20.1052590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func 2025-09-07T08:12:20.1052649Z return func(*args, **kwargs) 2025-09-07T08:12:20.1052927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 477, in forward 2025-09-07T08:12:20.1053045Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T08:12:20.1053323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 413, in forward 2025-09-07T08:12:20.1053399Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.1053403Z 2025-09-07T08:12:20.1053489Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1053678Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1053769Z return mod(**inputs) 2025-09-07T08:12:20.1054079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1054139Z outputs = self.roberta( 2025-09-07T08:12:20.1054407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1054474Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1054745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1054812Z layer_outputs = layer_module( 2025-09-07T08:12:20.1055025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1055090Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1055367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.1055474Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.1055743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.1055855Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.1056166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.1056282Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.1056554Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 493, in forward 2025-09-07T08:12:20.1056633Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.1056636Z 2025-09-07T08:12:20.1056727Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1056919Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1056977Z return mod(**inputs) 2025-09-07T08:12:20.1057252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1057318Z outputs = self.roberta( 2025-09-07T08:12:20.1057642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1057712Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1057983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1058044Z layer_outputs = layer_module( 2025-09-07T08:12:20.1058262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1058329Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1058611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.1058681Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.1058940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.1059004Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.1059305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 578, in feed_forward_chunk 2025-09-07T08:12:20.1059416Z intermediate_output = self.intermediate(attention_output) 2025-09-07T08:12:20.1059686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 494, in forward 2025-09-07T08:12:20.1059791Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T08:12:20.1060071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:12:20.1060135Z return self.act(input) 2025-09-07T08:12:20.1060141Z 2025-09-07T08:12:20.1060232Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1060420Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1060482Z return mod(**inputs) 2025-09-07T08:12:20.1060756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1094, in forward 2025-09-07T08:12:20.1060822Z outputs = self.roberta( 2025-09-07T08:12:20.1061091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 881, in forward 2025-09-07T08:12:20.1061155Z encoder_outputs = self.encoder( 2025-09-07T08:12:20.1061461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 632, in forward 2025-09-07T08:12:20.1061522Z layer_outputs = layer_module( 2025-09-07T08:12:20.1061736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:20.1061829Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:20.1062104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 570, in forward 2025-09-07T08:12:20.1062181Z layer_output = apply_chunking_to_forward( 2025-09-07T08:12:20.1062436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T08:12:20.1062507Z return forward_fn(*input_tensors) 2025-09-07T08:12:20.1062806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 579, in feed_forward_chunk 2025-09-07T08:12:20.1062937Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T08:12:20.1063206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 507, in forward 2025-09-07T08:12:20.1063276Z hidden_states = self.dense(hidden_states) 2025-09-07T08:12:20.1063279Z 2025-09-07T08:12:20.1063370Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1063553Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1063611Z return mod(**inputs) 2025-09-07T08:12:20.1063883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1108, in forward 2025-09-07T08:12:20.1063970Z prediction_scores = self.lm_head(sequence_output) 2025-09-07T08:12:20.1064242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1143, in forward 2025-09-07T08:12:20.1064306Z x = self.dense(features) 2025-09-07T08:12:20.1064310Z 2025-09-07T08:12:20.1064406Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:20.1064589Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:20.1064645Z return mod(**inputs) 2025-09-07T08:12:20.1064915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1108, in forward 2025-09-07T08:12:20.1064999Z prediction_scores = self.lm_head(sequence_output) 2025-09-07T08:12:20.1065279Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/xlm_roberta/modeling_xlm_roberta.py", line 1148, in forward 2025-09-07T08:12:20.1065337Z x = self.decoder(x) 2025-09-07T08:12:20.1065340Z 2025-09-07T08:12:31.2481510Z pass 2025-09-07T08:12:31.2481909Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:12:34.7636828Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:12:34.7637693Z import pynvml # type: ignore[import] 2025-09-07T08:12:36.7731013Z 2025-09-07T08:12:39.6209580Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:12:39.6210127Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:12:39.6429659Z cpu eval hf_T5 2025-09-07T08:12:40.6466727Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:12:41.1044304Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:12:41.5519672Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:12:53.3577254Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.3577975Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3578384Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3578736Z return mod(**inputs) 2025-09-07T08:12:53.3579215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.3587379Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.3587853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3588269Z layer_outputs = layer_module( 2025-09-07T08:12:53.3588638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3589013Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3589412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3589833Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3590216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3590597Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3590965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 546, in forward 2025-09-07T08:12:53.3591344Z position_bias = position_bias + causal_mask 2025-09-07T08:12:53.3591487Z 2025-09-07T08:12:53.3591594Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3591967Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3592297Z return mod(**inputs) 2025-09-07T08:12:53.3592648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.3593031Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.3593390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3593760Z layer_outputs = layer_module( 2025-09-07T08:12:53.3594114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3594496Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3594868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3595245Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3595620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:12:53.3596028Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.3596542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.3596986Z return self.weight * hidden_states 2025-09-07T08:12:53.3597126Z 2025-09-07T08:12:53.3597232Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3597603Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3597932Z return mod(**inputs) 2025-09-07T08:12:53.3598290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.3598664Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.3599034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3599407Z layer_outputs = layer_module( 2025-09-07T08:12:53.3599759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3600195Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3600587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3600962Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3601360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3601743Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3602122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.3602493Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.3602629Z 2025-09-07T08:12:53.3602730Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3603246Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3603576Z return mod(**inputs) 2025-09-07T08:12:53.3603932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.3604314Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.3604682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3605054Z layer_outputs = layer_module( 2025-09-07T08:12:53.3605539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3605894Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3606260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3606640Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3607003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3607379Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3607751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.3608129Z key_states = self.k(current_states) 2025-09-07T08:12:53.3608258Z 2025-09-07T08:12:53.3608361Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3608716Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3609029Z return mod(**inputs) 2025-09-07T08:12:53.3609387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.3609758Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.3610118Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3610546Z layer_outputs = layer_module( 2025-09-07T08:12:53.3610958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3611318Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3611680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3612053Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3612411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3612783Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3613147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.3613565Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.3613748Z 2025-09-07T08:12:53.3613849Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3614244Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3614572Z return mod(**inputs) 2025-09-07T08:12:53.3614919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.3615331Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.3615696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3616061Z layer_outputs = layer_module( 2025-09-07T08:12:53.3616403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3616761Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3617137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3617510Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3617942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3618329Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3618696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.3619113Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.3619290Z 2025-09-07T08:12:53.3619392Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3619746Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3620066Z return mod(**inputs) 2025-09-07T08:12:53.3620410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.3620770Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.3621133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3621500Z layer_outputs = layer_module( 2025-09-07T08:12:53.3621844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3622202Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3622559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3622927Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3623294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3623668Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3624036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.3624486Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.3624710Z 2025-09-07T08:12:53.3624809Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3625168Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3625485Z return mod(**inputs) 2025-09-07T08:12:53.3625840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.3626207Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.3626566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3626931Z layer_outputs = layer_module( 2025-09-07T08:12:53.3627267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3627618Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3628033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3628441Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3628836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3629222Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3629579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.3630027Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.3630241Z 2025-09-07T08:12:53.3630337Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3630690Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3631004Z return mod(**inputs) 2025-09-07T08:12:53.3631340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.3631707Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.3632067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3632433Z layer_outputs = layer_module( 2025-09-07T08:12:53.3632759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3633119Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3633481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3633864Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3634235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3634622Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3634988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.3635357Z value_states = self.v(current_states) 2025-09-07T08:12:53.3635492Z 2025-09-07T08:12:53.3635594Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3635939Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3636244Z return mod(**inputs) 2025-09-07T08:12:53.3636576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.3636946Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.3637297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3637659Z layer_outputs = layer_module( 2025-09-07T08:12:53.3638030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3638434Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3638797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3639166Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3639528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3639897Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3640261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.3640662Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.3640821Z 2025-09-07T08:12:53.3640916Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3641262Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3641626Z return mod(**inputs) 2025-09-07T08:12:53.3641963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.3642356Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.3642743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3643131Z layer_outputs = layer_module( 2025-09-07T08:12:53.3643478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3643842Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3644198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3644567Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3644936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3645309Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3645667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.3646067Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.3646225Z 2025-09-07T08:12:53.3646339Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3646681Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3646997Z return mod(**inputs) 2025-09-07T08:12:53.3647318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.3647689Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.3648045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3648411Z layer_outputs = layer_module( 2025-09-07T08:12:53.3648748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3649094Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3649459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3649830Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3650190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3650558Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3650917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.3651303Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.3651502Z 2025-09-07T08:12:53.3651593Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3651974Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3652294Z return mod(**inputs) 2025-09-07T08:12:53.3652637Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.3653001Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.3653361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3653730Z layer_outputs = layer_module( 2025-09-07T08:12:53.3654053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3654413Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3654769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3655182Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3655544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3655911Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3656303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.3656705Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.3656857Z 2025-09-07T08:12:53.3656954Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3657302Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3657645Z return mod(**inputs) 2025-09-07T08:12:53.3657978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.3658350Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.3658710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3659070Z layer_outputs = layer_module( 2025-09-07T08:12:53.3659402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3659755Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3660112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3660482Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3660842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3661207Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3661573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.3661941Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.3662069Z 2025-09-07T08:12:53.3662172Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3662508Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3662832Z return mod(**inputs) 2025-09-07T08:12:53.3663165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.3663544Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.3663892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3664261Z layer_outputs = layer_module( 2025-09-07T08:12:53.3664590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3665033Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3665440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.3665819Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.3666189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.3666571Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.3666935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.3667302Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.3667442Z 2025-09-07T08:12:53.3667542Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3667898Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3668229Z return mod(**inputs) 2025-09-07T08:12:53.3668607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3668983Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3669343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3669742Z layer_outputs = layer_module( 2025-09-07T08:12:53.3670087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3670447Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3670800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3671189Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3671552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3671925Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3672283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.3672651Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.3672788Z 2025-09-07T08:12:53.3672887Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3673233Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3673547Z return mod(**inputs) 2025-09-07T08:12:53.3673879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3674244Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3674601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3674964Z layer_outputs = layer_module( 2025-09-07T08:12:53.3675298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3675654Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3676018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3676393Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3676755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3677120Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3677486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.3677851Z key_states = self.k(current_states) 2025-09-07T08:12:53.3677979Z 2025-09-07T08:12:53.3678082Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3678427Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3678815Z return mod(**inputs) 2025-09-07T08:12:53.3679152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3679520Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3679875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3680234Z layer_outputs = layer_module( 2025-09-07T08:12:53.3680566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3680920Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3681286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3681659Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3682014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3682431Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3682794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.3684293Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.3684485Z 2025-09-07T08:12:53.3684592Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3684943Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3685261Z return mod(**inputs) 2025-09-07T08:12:53.3685599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3685969Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3686315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3686683Z layer_outputs = layer_module( 2025-09-07T08:12:53.3687018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3687373Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3687733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3688094Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3688453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3688825Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3689186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.3689588Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.3689772Z 2025-09-07T08:12:53.3689868Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3690213Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3690528Z return mod(**inputs) 2025-09-07T08:12:53.3690861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3691221Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3691574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3691939Z layer_outputs = layer_module( 2025-09-07T08:12:53.3692272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3692624Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3692976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3693473Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3693854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3694230Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3694598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.3695010Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.3695193Z 2025-09-07T08:12:53.3695287Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3695627Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3695938Z return mod(**inputs) 2025-09-07T08:12:53.3696262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3696632Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3697022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3697394Z layer_outputs = layer_module( 2025-09-07T08:12:53.3697810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3698166Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3698522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3698889Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3699264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3699628Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3699989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.3700444Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.3700651Z 2025-09-07T08:12:53.3700755Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3701102Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3701410Z return mod(**inputs) 2025-09-07T08:12:53.3701742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3702106Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3702456Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3702818Z layer_outputs = layer_module( 2025-09-07T08:12:53.3703320Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3703683Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3704052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3704419Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3704775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3705143Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3705504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.3705943Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.3706149Z 2025-09-07T08:12:53.3706250Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3706585Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3706980Z return mod(**inputs) 2025-09-07T08:12:53.3707362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3707726Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3708074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3708437Z layer_outputs = layer_module( 2025-09-07T08:12:53.3708767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3709122Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3709477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3709842Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3710199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3710673Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3711037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.3711398Z value_states = self.v(current_states) 2025-09-07T08:12:53.3711571Z 2025-09-07T08:12:53.3711669Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3712014Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3712328Z return mod(**inputs) 2025-09-07T08:12:53.3712662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3713024Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3713376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3713739Z layer_outputs = layer_module( 2025-09-07T08:12:53.3714073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3714430Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3714782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3715154Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3715511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3715877Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3716235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.3716626Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.3716792Z 2025-09-07T08:12:53.3716883Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3717229Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3717544Z return mod(**inputs) 2025-09-07T08:12:53.3717882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3718265Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3718634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3719003Z layer_outputs = layer_module( 2025-09-07T08:12:53.3719352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3719709Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3720080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3720500Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3720911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3721281Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3721650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.3722049Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.3722209Z 2025-09-07T08:12:53.3722317Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3722671Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3722987Z return mod(**inputs) 2025-09-07T08:12:53.3723329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3723699Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3724090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3724460Z layer_outputs = layer_module( 2025-09-07T08:12:53.3724798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3725187Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3725559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3725932Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3726287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3726655Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3727018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.3727414Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.3727570Z 2025-09-07T08:12:53.3727677Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3728017Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3728336Z return mod(**inputs) 2025-09-07T08:12:53.3728671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3729039Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3729390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3729749Z layer_outputs = layer_module( 2025-09-07T08:12:53.3730086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3730443Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3730806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3731171Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3731537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3731912Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3732274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.3732663Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.3732818Z 2025-09-07T08:12:53.3732913Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3733260Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3733573Z return mod(**inputs) 2025-09-07T08:12:53.3733913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3734358Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3734727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3735099Z layer_outputs = layer_module( 2025-09-07T08:12:53.3735447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3735813Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3736184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3736568Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3736956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3737332Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3737800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.3738187Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.3738319Z 2025-09-07T08:12:53.3738397Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.3738667Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3739025Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3739339Z return mod(**inputs) 2025-09-07T08:12:53.3739687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3740061Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3740420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3740785Z layer_outputs = layer_module( 2025-09-07T08:12:53.3741128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3741491Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3741850Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.3742237Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.3742612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:12:53.3743002Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.3743384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.3743753Z return self.weight * hidden_states 2025-09-07T08:12:53.3743880Z 2025-09-07T08:12:53.3743988Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3744336Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3744664Z return mod(**inputs) 2025-09-07T08:12:53.3745008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3745380Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3745738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3746106Z layer_outputs = layer_module( 2025-09-07T08:12:53.3746443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3746801Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3747168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.3747542Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.3747981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.3748431Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.3748848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:12:53.3749220Z hidden_states = self.wi(hidden_states) 2025-09-07T08:12:53.3749350Z 2025-09-07T08:12:53.3749450Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3749802Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3750123Z return mod(**inputs) 2025-09-07T08:12:53.3750462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3750832Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3751186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3751595Z layer_outputs = layer_module( 2025-09-07T08:12:53.3751943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3752334Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3752695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.3753081Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.3753458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.3753868Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.3754265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:12:53.3754631Z hidden_states = self.act(hidden_states) 2025-09-07T08:12:53.3754768Z 2025-09-07T08:12:53.3754864Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3755213Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3755528Z return mod(**inputs) 2025-09-07T08:12:53.3755857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3756223Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3756576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3756938Z layer_outputs = layer_module( 2025-09-07T08:12:53.3757271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3757619Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3757982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.3758363Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.3758740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.3759143Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.3759533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:12:53.3759905Z hidden_states = self.wo(hidden_states) 2025-09-07T08:12:53.3760039Z 2025-09-07T08:12:53.3760116Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.3760347Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3760688Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3761000Z return mod(**inputs) 2025-09-07T08:12:53.3761379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3761788Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3762140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3762502Z layer_outputs = layer_module( 2025-09-07T08:12:53.3762839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3763192Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3763553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3763917Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3764281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:12:53.3764696Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.3765122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.3765494Z return self.weight * hidden_states 2025-09-07T08:12:53.3765621Z 2025-09-07T08:12:53.3765748Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3766097Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3766414Z return mod(**inputs) 2025-09-07T08:12:53.3766750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3767118Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3767466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3767832Z layer_outputs = layer_module( 2025-09-07T08:12:53.3768171Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3768529Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3768885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3769259Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3769621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3769995Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3770358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.3770720Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.3770854Z 2025-09-07T08:12:53.3770948Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3771290Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3771611Z return mod(**inputs) 2025-09-07T08:12:53.3771948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3772307Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3772664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3773028Z layer_outputs = layer_module( 2025-09-07T08:12:53.3773363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3773711Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3774068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3774439Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3774844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3775253Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3775609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.3775988Z key_states = self.k(current_states) 2025-09-07T08:12:53.3776124Z 2025-09-07T08:12:53.3776221Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3776582Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3776888Z return mod(**inputs) 2025-09-07T08:12:53.3777238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3777670Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3778032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3778404Z layer_outputs = layer_module( 2025-09-07T08:12:53.3778764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3779121Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3779510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3779890Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3780254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3780616Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3780981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.3781397Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.3781583Z 2025-09-07T08:12:53.3781685Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3782028Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3782344Z return mod(**inputs) 2025-09-07T08:12:53.3782679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3783050Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3783403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3783758Z layer_outputs = layer_module( 2025-09-07T08:12:53.3784090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3784443Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3784802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3785174Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3785530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3785900Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3786263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.3786674Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.3786849Z 2025-09-07T08:12:53.3786942Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3787286Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3787604Z return mod(**inputs) 2025-09-07T08:12:53.3787935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3788357Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3788741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3789106Z layer_outputs = layer_module( 2025-09-07T08:12:53.3789435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3789794Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3790148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3790521Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3790878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3791243Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3791605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.3792068Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.3792251Z 2025-09-07T08:12:53.3792343Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3792731Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3793055Z return mod(**inputs) 2025-09-07T08:12:53.3793392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3793754Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3794111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3794474Z layer_outputs = layer_module( 2025-09-07T08:12:53.3794805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3795154Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3795517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3795888Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3796253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3796621Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3796977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.3797429Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.3797640Z 2025-09-07T08:12:53.3797734Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3798075Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3798391Z return mod(**inputs) 2025-09-07T08:12:53.3798721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3799087Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3799444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3799809Z layer_outputs = layer_module( 2025-09-07T08:12:53.3800135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3800490Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3800846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3801216Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3801575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3802023Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3802382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.3802818Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.3803019Z 2025-09-07T08:12:53.3803275Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3803620Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3803928Z return mod(**inputs) 2025-09-07T08:12:53.3804264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3804630Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3804987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3805349Z layer_outputs = layer_module( 2025-09-07T08:12:53.3805759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3806124Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3806542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3806927Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3807290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3807662Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3808026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.3808392Z value_states = self.v(current_states) 2025-09-07T08:12:53.3808523Z 2025-09-07T08:12:53.3808627Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3808975Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3809293Z return mod(**inputs) 2025-09-07T08:12:53.3809625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3809996Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3810343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3810703Z layer_outputs = layer_module( 2025-09-07T08:12:53.3811034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3811386Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3811746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3812115Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3812483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3812848Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3813212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.3813606Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.3813775Z 2025-09-07T08:12:53.3813870Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3814213Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3814526Z return mod(**inputs) 2025-09-07T08:12:53.3814855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3815214Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3815626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3816044Z layer_outputs = layer_module( 2025-09-07T08:12:53.3816386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3816748Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3817104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3817476Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3817898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3818270Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3818622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.3819021Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.3819184Z 2025-09-07T08:12:53.3819315Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3819659Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3820009Z return mod(**inputs) 2025-09-07T08:12:53.3820335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3820704Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3821058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3821421Z layer_outputs = layer_module( 2025-09-07T08:12:53.3821748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3822101Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3822468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3822841Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3823200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3823565Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3823933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.3824324Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.3824478Z 2025-09-07T08:12:53.3824583Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3824927Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3825232Z return mod(**inputs) 2025-09-07T08:12:53.3825565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3825961Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3826311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3826668Z layer_outputs = layer_module( 2025-09-07T08:12:53.3826999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3827353Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3827714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3828080Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3828438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3828827Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3829248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.3829688Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.3829848Z 2025-09-07T08:12:53.3829967Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3830319Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3830646Z return mod(**inputs) 2025-09-07T08:12:53.3830992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3831364Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3831720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3832088Z layer_outputs = layer_module( 2025-09-07T08:12:53.3843629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3844158Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3844565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3844999Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3845378Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3845796Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3846172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.3846574Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.3846702Z 2025-09-07T08:12:53.3846813Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3847179Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3847504Z return mod(**inputs) 2025-09-07T08:12:53.3847874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3848245Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3848603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3848959Z layer_outputs = layer_module( 2025-09-07T08:12:53.3849327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3849679Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3850037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.3850431Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.3850808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:12:53.3851200Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.3851574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.3851931Z return self.weight * hidden_states 2025-09-07T08:12:53.3852082Z 2025-09-07T08:12:53.3852180Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3852526Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3852855Z return mod(**inputs) 2025-09-07T08:12:53.3853192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3853571Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3853930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3854343Z layer_outputs = layer_module( 2025-09-07T08:12:53.3854732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3855083Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3855455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.3855838Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.3856235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.3856657Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.3857050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:12:53.3857432Z hidden_states = self.wi(hidden_states) 2025-09-07T08:12:53.3857650Z 2025-09-07T08:12:53.3857750Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3858163Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3858484Z return mod(**inputs) 2025-09-07T08:12:53.3858856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3859231Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3859589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3859951Z layer_outputs = layer_module( 2025-09-07T08:12:53.3860292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3860646Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3861012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.3861396Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.3861773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.3862184Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.3862583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:12:53.3862957Z hidden_states = self.act(hidden_states) 2025-09-07T08:12:53.3863087Z 2025-09-07T08:12:53.3863192Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3863537Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3863862Z return mod(**inputs) 2025-09-07T08:12:53.3864214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3864583Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3864947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3865317Z layer_outputs = layer_module( 2025-09-07T08:12:53.3865668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3866015Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3866377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.3866754Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.3867120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.3867531Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.3867920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:12:53.3868372Z hidden_states = self.wo(hidden_states) 2025-09-07T08:12:53.3868502Z 2025-09-07T08:12:53.3868577Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.3868812Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3869155Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3869477Z return mod(**inputs) 2025-09-07T08:12:53.3869806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3870167Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3870537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3870909Z layer_outputs = layer_module( 2025-09-07T08:12:53.3871234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3871580Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3872171Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3872541Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3872937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:12:53.3873334Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.3873714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.3874084Z return self.weight * hidden_states 2025-09-07T08:12:53.3874210Z 2025-09-07T08:12:53.3874304Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3874651Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3874961Z return mod(**inputs) 2025-09-07T08:12:53.3875292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3875659Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3876009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3876374Z layer_outputs = layer_module( 2025-09-07T08:12:53.3876697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3877068Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3877413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3877771Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3878168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3878532Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3878898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.3879257Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.3879384Z 2025-09-07T08:12:53.3879479Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3879812Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3880116Z return mod(**inputs) 2025-09-07T08:12:53.3880438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3880794Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3881143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3881548Z layer_outputs = layer_module( 2025-09-07T08:12:53.3881918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3882262Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3882622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3882989Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3883334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3883706Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3884056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.3884411Z key_states = self.k(current_states) 2025-09-07T08:12:53.3884541Z 2025-09-07T08:12:53.3884634Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3885007Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3885325Z return mod(**inputs) 2025-09-07T08:12:53.3885648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3886039Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3886393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3886754Z layer_outputs = layer_module( 2025-09-07T08:12:53.3887085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3887447Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3887808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3888183Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3888554Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3888922Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3889283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.3889690Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.3889867Z 2025-09-07T08:12:53.3889963Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3890314Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3890632Z return mod(**inputs) 2025-09-07T08:12:53.3890969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3891326Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3891681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3892041Z layer_outputs = layer_module( 2025-09-07T08:12:53.3892369Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3892715Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3893075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3893451Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3893821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3894180Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3894531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.3894991Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.3895197Z 2025-09-07T08:12:53.3895294Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3895634Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3895941Z return mod(**inputs) 2025-09-07T08:12:53.3896269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3896625Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3896969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3897326Z layer_outputs = layer_module( 2025-09-07T08:12:53.3897729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3898079Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3898484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3898857Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3899221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3899616Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3899969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.3900375Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.3900552Z 2025-09-07T08:12:53.3900646Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3900984Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3901297Z return mod(**inputs) 2025-09-07T08:12:53.3901623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3901986Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3902327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3902681Z layer_outputs = layer_module( 2025-09-07T08:12:53.3903016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3903575Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3903939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3904305Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3904663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3905021Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3905384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.3905821Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.3906024Z 2025-09-07T08:12:53.3906125Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3906471Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3906777Z return mod(**inputs) 2025-09-07T08:12:53.3907105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3907465Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3907810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3908162Z layer_outputs = layer_module( 2025-09-07T08:12:53.3908590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3909004Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3909354Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3909716Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3910067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3910435Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3910789Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.3911225Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.3911426Z 2025-09-07T08:12:53.3911525Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3911869Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3912247Z return mod(**inputs) 2025-09-07T08:12:53.3912579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3912931Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3913323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3913686Z layer_outputs = layer_module( 2025-09-07T08:12:53.3914021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3914371Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3914723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3915080Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3915433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3915791Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3916146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.3916509Z value_states = self.v(current_states) 2025-09-07T08:12:53.3916637Z 2025-09-07T08:12:53.3916728Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3917062Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3917364Z return mod(**inputs) 2025-09-07T08:12:53.3917685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3918037Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3918376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3918735Z layer_outputs = layer_module( 2025-09-07T08:12:53.3919059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3919407Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3919761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3920122Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3920475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3920838Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3921194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.3921587Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.3921791Z 2025-09-07T08:12:53.3921882Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3922302Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3922606Z return mod(**inputs) 2025-09-07T08:12:53.3922926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3923278Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3923621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3923975Z layer_outputs = layer_module( 2025-09-07T08:12:53.3924303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3924642Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3924994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3925390Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3925740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3926102Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3926487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.3926878Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.3927030Z 2025-09-07T08:12:53.3927126Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3927459Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3927761Z return mod(**inputs) 2025-09-07T08:12:53.3928080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3928436Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3928777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3929127Z layer_outputs = layer_module( 2025-09-07T08:12:53.3929448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3929791Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3930146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3930510Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3930858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3931215Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3931572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.3931962Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.3932109Z 2025-09-07T08:12:53.3932201Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3932527Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3932830Z return mod(**inputs) 2025-09-07T08:12:53.3933149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3933506Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3933851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3934202Z layer_outputs = layer_module( 2025-09-07T08:12:53.3934526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3934917Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3935348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3935705Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3936055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3936410Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3936758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.3937137Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.3937285Z 2025-09-07T08:12:53.3937374Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3937753Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3938057Z return mod(**inputs) 2025-09-07T08:12:53.3938414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3938779Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3939148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3939509Z layer_outputs = layer_module( 2025-09-07T08:12:53.3939840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3940180Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3940529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3940890Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3941240Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3941603Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3941958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.3942314Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.3942438Z 2025-09-07T08:12:53.3942512Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.3942727Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3943065Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3943377Z return mod(**inputs) 2025-09-07T08:12:53.3943711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3944072Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3944415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3944772Z layer_outputs = layer_module( 2025-09-07T08:12:53.3945099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3945448Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3945800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.3946172Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.3946535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:12:53.3946913Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.3947300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.3947667Z return self.weight * hidden_states 2025-09-07T08:12:53.3947794Z 2025-09-07T08:12:53.3947956Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3948338Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3948653Z return mod(**inputs) 2025-09-07T08:12:53.3948990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3949352Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3949704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3950071Z layer_outputs = layer_module( 2025-09-07T08:12:53.3950408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3950769Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3951126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.3951497Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.3951901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.3952306Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.3952733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:12:53.3953101Z hidden_states = self.wi(hidden_states) 2025-09-07T08:12:53.3953230Z 2025-09-07T08:12:53.3953328Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3953674Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3953981Z return mod(**inputs) 2025-09-07T08:12:53.3954308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3954667Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3955013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3955370Z layer_outputs = layer_module( 2025-09-07T08:12:53.3955698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3956042Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3956389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.3956754Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.3957112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.3957507Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.3957908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:12:53.3958277Z hidden_states = self.act(hidden_states) 2025-09-07T08:12:53.3958414Z 2025-09-07T08:12:53.3958510Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3958852Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3959169Z return mod(**inputs) 2025-09-07T08:12:53.3959501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3959860Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3960223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3960585Z layer_outputs = layer_module( 2025-09-07T08:12:53.3960916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3961260Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3961684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.3962093Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.3962465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.3962861Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.3963249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:12:53.3963613Z hidden_states = self.wo(hidden_states) 2025-09-07T08:12:53.3963617Z 2025-09-07T08:12:53.3963695Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.3963788Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3963977Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3964046Z return mod(**inputs) 2025-09-07T08:12:53.3964314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3964388Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3964650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3964713Z layer_outputs = layer_module( 2025-09-07T08:12:53.3964936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3965006Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3965242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3965315Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3965542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:12:53.3965651Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.3965878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.3965952Z return self.weight * hidden_states 2025-09-07T08:12:53.3965957Z 2025-09-07T08:12:53.3966048Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3966241Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3966299Z return mod(**inputs) 2025-09-07T08:12:53.3966530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3966600Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3966832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3966902Z layer_outputs = layer_module( 2025-09-07T08:12:53.3967120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3967189Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3967424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3967496Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3967728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3967804Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3968036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.3968108Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.3968112Z 2025-09-07T08:12:53.3968201Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3968429Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3968517Z return mod(**inputs) 2025-09-07T08:12:53.3968756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3968822Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3969054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3969123Z layer_outputs = layer_module( 2025-09-07T08:12:53.3969340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3969416Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3969644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3969717Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3969987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3970061Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3970325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.3970396Z key_states = self.k(current_states) 2025-09-07T08:12:53.3970400Z 2025-09-07T08:12:53.3970493Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3970679Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3970739Z return mod(**inputs) 2025-09-07T08:12:53.3970978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3971041Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3971281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3971346Z layer_outputs = layer_module( 2025-09-07T08:12:53.3971559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3971633Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3971859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3971937Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3972166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3972237Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3972471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.3972589Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.3972595Z 2025-09-07T08:12:53.3972694Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3972877Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3972940Z return mod(**inputs) 2025-09-07T08:12:53.3973171Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3973236Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3973472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3973534Z layer_outputs = layer_module( 2025-09-07T08:12:53.3973751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3973818Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3974099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3974212Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3974443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3974524Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3974751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.3974866Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.3974876Z 2025-09-07T08:12:53.3974973Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3975165Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3975229Z return mod(**inputs) 2025-09-07T08:12:53.3975464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3975568Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3975805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3975872Z layer_outputs = layer_module( 2025-09-07T08:12:53.3976125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3976197Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3976435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3976506Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3976732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3976813Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3977050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.3977171Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.3977175Z 2025-09-07T08:12:53.3977264Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3977456Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3977514Z return mod(**inputs) 2025-09-07T08:12:53.3977838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3977909Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3978141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3978209Z layer_outputs = layer_module( 2025-09-07T08:12:53.3978426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3978501Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3978734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3978806Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3979039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3979110Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3979339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.3979494Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.3979497Z 2025-09-07T08:12:53.3979588Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3979780Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3979904Z return mod(**inputs) 2025-09-07T08:12:53.3980145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3980208Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3980442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3980510Z layer_outputs = layer_module( 2025-09-07T08:12:53.3980724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3980797Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3981024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3981095Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3981326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3981430Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3981669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.3981849Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.3981853Z 2025-09-07T08:12:53.3981951Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3982136Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3982194Z return mod(**inputs) 2025-09-07T08:12:53.3982431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3982493Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3982733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3982800Z layer_outputs = layer_module( 2025-09-07T08:12:53.3983012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3983086Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3983317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3983396Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3983623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3983694Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3983923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.3983990Z value_states = self.v(current_states) 2025-09-07T08:12:53.3983995Z 2025-09-07T08:12:53.3984084Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3984267Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3984323Z return mod(**inputs) 2025-09-07T08:12:53.3984553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3984614Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3984844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3984902Z layer_outputs = layer_module( 2025-09-07T08:12:53.3985117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3985180Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3985405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3985538Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3985764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3985834Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3986061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.3986160Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.3986166Z 2025-09-07T08:12:53.3986254Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3986436Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3986493Z return mod(**inputs) 2025-09-07T08:12:53.3986722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3986787Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3987059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3987119Z layer_outputs = layer_module( 2025-09-07T08:12:53.3987374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3987439Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3987667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3987734Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3987959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3988028Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3988252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.3988350Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.3988353Z 2025-09-07T08:12:53.3988439Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3988623Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3988678Z return mod(**inputs) 2025-09-07T08:12:53.3988905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3988968Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3989197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3989257Z layer_outputs = layer_module( 2025-09-07T08:12:53.3989467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3989533Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3989763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3989829Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3990058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3990125Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3990349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.3990439Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.3990443Z 2025-09-07T08:12:53.3990529Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3990710Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3990800Z return mod(**inputs) 2025-09-07T08:12:53.3991032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3991124Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3991354Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3991418Z layer_outputs = layer_module( 2025-09-07T08:12:53.3991629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3991695Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3991920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3991987Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3992216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3992285Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3992542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.3992633Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.3992636Z 2025-09-07T08:12:53.3992753Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3992936Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3992990Z return mod(**inputs) 2025-09-07T08:12:53.3993219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3993279Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3993509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3993568Z layer_outputs = layer_module( 2025-09-07T08:12:53.3993782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3993851Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3994076Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3994146Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3994370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.3994438Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.3994665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.3994731Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.3994734Z 2025-09-07T08:12:53.3994825Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3995009Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3995070Z return mod(**inputs) 2025-09-07T08:12:53.3995303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3995368Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3995601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3995662Z layer_outputs = layer_module( 2025-09-07T08:12:53.3995877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3995946Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3996172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.3996244Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.3996508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:12:53.3996672Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:12:53.3996675Z 2025-09-07T08:12:53.3996746Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.3996839Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3997027Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3997082Z return mod(**inputs) 2025-09-07T08:12:53.3997315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3997377Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3997610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3997671Z layer_outputs = layer_module( 2025-09-07T08:12:53.3997918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.3997989Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.3998246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.3998329Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.3998555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:12:53.3998641Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.3998871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.3998936Z return self.weight * hidden_states 2025-09-07T08:12:53.3998939Z 2025-09-07T08:12:53.3999033Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.3999221Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.3999279Z return mod(**inputs) 2025-09-07T08:12:53.3999512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.3999577Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.3999811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.3999871Z layer_outputs = layer_module( 2025-09-07T08:12:53.4000083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4000149Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4000374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4000454Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4000685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4000795Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4001023Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:12:53.4001094Z hidden_states = self.wi(hidden_states) 2025-09-07T08:12:53.4001097Z 2025-09-07T08:12:53.4001191Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4001375Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4001434Z return mod(**inputs) 2025-09-07T08:12:53.4001661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4001726Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4001988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4002081Z layer_outputs = layer_module( 2025-09-07T08:12:53.4002298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4002367Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4002599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4002676Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4002901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4003006Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4003468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:12:53.4003544Z hidden_states = self.act(hidden_states) 2025-09-07T08:12:53.4003549Z 2025-09-07T08:12:53.4003718Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4003910Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4003971Z return mod(**inputs) 2025-09-07T08:12:53.4004261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4004330Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4004562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4004626Z layer_outputs = layer_module( 2025-09-07T08:12:53.4004842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4004910Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4005149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4005233Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4005462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4005568Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4005796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:12:53.4005869Z hidden_states = self.wo(hidden_states) 2025-09-07T08:12:53.4005872Z 2025-09-07T08:12:53.4005944Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4006038Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4006225Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4006281Z return mod(**inputs) 2025-09-07T08:12:53.4006522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4006589Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4006826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4006891Z layer_outputs = layer_module( 2025-09-07T08:12:53.4007113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4007186Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4007415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4007494Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4007721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:12:53.4007872Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4008155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4008227Z return self.weight * hidden_states 2025-09-07T08:12:53.4008230Z 2025-09-07T08:12:53.4008335Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4008526Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4008594Z return mod(**inputs) 2025-09-07T08:12:53.4008830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4008897Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4009136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4009202Z layer_outputs = layer_module( 2025-09-07T08:12:53.4009458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4009531Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4009769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4009874Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4010107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4010189Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4010415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.4010493Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.4010496Z 2025-09-07T08:12:53.4010588Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4010780Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4010853Z return mod(**inputs) 2025-09-07T08:12:53.4011088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4011158Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4011388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4011456Z layer_outputs = layer_module( 2025-09-07T08:12:53.4011672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4011740Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4011975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4012050Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4012287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4012366Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4012595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.4012671Z key_states = self.k(current_states) 2025-09-07T08:12:53.4012675Z 2025-09-07T08:12:53.4012762Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4012954Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4013011Z return mod(**inputs) 2025-09-07T08:12:53.4013243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4013308Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4013540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4013669Z layer_outputs = layer_module( 2025-09-07T08:12:53.4013885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4013956Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4014186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4014258Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4014491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4014561Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4014790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4014914Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4014919Z 2025-09-07T08:12:53.4015010Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4015228Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4015284Z return mod(**inputs) 2025-09-07T08:12:53.4015553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4015616Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4015846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4015932Z layer_outputs = layer_module( 2025-09-07T08:12:53.4016156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4016228Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4016454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4016532Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4016757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4016827Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4017076Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4017186Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4017190Z 2025-09-07T08:12:53.4017283Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4017471Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4017536Z return mod(**inputs) 2025-09-07T08:12:53.4017830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4017896Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4018143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4018203Z layer_outputs = layer_module( 2025-09-07T08:12:53.4018417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4018486Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4018715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4018795Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4019020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4019089Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4019312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4019487Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4019490Z 2025-09-07T08:12:53.4019583Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4019769Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4019828Z return mod(**inputs) 2025-09-07T08:12:53.4020057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4020124Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4020363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4020426Z layer_outputs = layer_module( 2025-09-07T08:12:53.4020641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4020709Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4020971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4021038Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4021301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4021375Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4021601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.4021765Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.4021768Z 2025-09-07T08:12:53.4021854Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4022039Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4022097Z return mod(**inputs) 2025-09-07T08:12:53.4022332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4022402Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4022630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4022689Z layer_outputs = layer_module( 2025-09-07T08:12:53.4022907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4022972Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4023211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4023279Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4023520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4023591Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4023820Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.4023963Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.4023966Z 2025-09-07T08:12:53.4024055Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4024241Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4024296Z return mod(**inputs) 2025-09-07T08:12:53.4024527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4024601Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4024833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4024934Z layer_outputs = layer_module( 2025-09-07T08:12:53.4025195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4025272Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4025504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4025583Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4025822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4025906Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4026138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.4026208Z value_states = self.v(current_states) 2025-09-07T08:12:53.4026211Z 2025-09-07T08:12:53.4026304Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4026526Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4026584Z return mod(**inputs) 2025-09-07T08:12:53.4026853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4026919Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4027151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4027232Z layer_outputs = layer_module( 2025-09-07T08:12:53.4027446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4027531Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4027758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4027833Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4028061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4028133Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4028365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4028466Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4028470Z 2025-09-07T08:12:53.4028564Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4028747Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4028803Z return mod(**inputs) 2025-09-07T08:12:53.4029051Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4029114Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4029355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4029429Z layer_outputs = layer_module( 2025-09-07T08:12:53.4029647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4029722Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4029952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4030028Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4030263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4030334Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4030563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4030688Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4030720Z 2025-09-07T08:12:53.4030815Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4031006Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4031067Z return mod(**inputs) 2025-09-07T08:12:53.4031297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4031359Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4031590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4031651Z layer_outputs = layer_module( 2025-09-07T08:12:53.4031882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4031946Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4032205Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4032283Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4032543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4032620Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4032845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4032938Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4033199Z 2025-09-07T08:12:53.4033306Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4033503Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4033568Z return mod(**inputs) 2025-09-07T08:12:53.4033805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4033898Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4034128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4034192Z layer_outputs = layer_module( 2025-09-07T08:12:53.4034402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4034466Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4034709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4034776Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4035008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4035082Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4035316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.4035415Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.4035419Z 2025-09-07T08:12:53.4035509Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4035700Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4035758Z return mod(**inputs) 2025-09-07T08:12:53.4035991Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4036059Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4036289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4036366Z layer_outputs = layer_module( 2025-09-07T08:12:53.4036580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4036710Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4036940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4037011Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4037251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4037320Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4037553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.4037620Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.4037624Z 2025-09-07T08:12:53.4037696Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4037792Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4037978Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4038071Z return mod(**inputs) 2025-09-07T08:12:53.4038307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4038403Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4038651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4038712Z layer_outputs = layer_module( 2025-09-07T08:12:53.4038933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4038999Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4039236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4039319Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4039549Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:12:53.4039640Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4039868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4039947Z return self.weight * hidden_states 2025-09-07T08:12:53.4039951Z 2025-09-07T08:12:53.4040039Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4040223Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4040284Z return mod(**inputs) 2025-09-07T08:12:53.4040513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4040579Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4040818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4040883Z layer_outputs = layer_module( 2025-09-07T08:12:53.4041097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4041165Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4041396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4041472Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4041708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4041815Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4042042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:12:53.4042182Z hidden_states = self.wi(hidden_states) 2025-09-07T08:12:53.4042213Z 2025-09-07T08:12:53.4042304Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4042492Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4042549Z return mod(**inputs) 2025-09-07T08:12:53.4042781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4042846Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4043077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4043145Z layer_outputs = layer_module( 2025-09-07T08:12:53.4043356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4043426Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4043652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4043765Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4043996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4044123Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4044367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:12:53.4044438Z hidden_states = self.act(hidden_states) 2025-09-07T08:12:53.4044442Z 2025-09-07T08:12:53.4044529Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4044717Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4044771Z return mod(**inputs) 2025-09-07T08:12:53.4045005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4045071Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4045302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4045366Z layer_outputs = layer_module( 2025-09-07T08:12:53.4045582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4045654Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4045883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4045964Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4046191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4046290Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4046524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:12:53.4046592Z hidden_states = self.wo(hidden_states) 2025-09-07T08:12:53.4046595Z 2025-09-07T08:12:53.4046670Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4046758Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4046940Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4047000Z return mod(**inputs) 2025-09-07T08:12:53.4047236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4047306Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4047536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4047598Z layer_outputs = layer_module( 2025-09-07T08:12:53.4047864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4047959Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4048194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4048264Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4048499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:12:53.4048596Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4048826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4048896Z return self.weight * hidden_states 2025-09-07T08:12:53.4048899Z 2025-09-07T08:12:53.4048985Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4049176Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4049263Z return mod(**inputs) 2025-09-07T08:12:53.4049499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4049567Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4049829Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4049898Z layer_outputs = layer_module( 2025-09-07T08:12:53.4050111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4050181Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4050409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4050478Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4050712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4050783Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4051015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.4051083Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.4051086Z 2025-09-07T08:12:53.4051173Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4051358Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4051416Z return mod(**inputs) 2025-09-07T08:12:53.4051651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4051711Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4051941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4052011Z layer_outputs = layer_module( 2025-09-07T08:12:53.4052220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4052289Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4052517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4052586Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4052810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4052878Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4053107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.4053178Z key_states = self.k(current_states) 2025-09-07T08:12:53.4053214Z 2025-09-07T08:12:53.4053303Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4053517Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4053573Z return mod(**inputs) 2025-09-07T08:12:53.4053808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4053869Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4054100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4054162Z layer_outputs = layer_module( 2025-09-07T08:12:53.4054375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4054443Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4054671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4054747Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4055005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4055080Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4055340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4055462Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4055465Z 2025-09-07T08:12:53.4055562Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4055747Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4055804Z return mod(**inputs) 2025-09-07T08:12:53.4056035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4056101Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4056336Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4056398Z layer_outputs = layer_module( 2025-09-07T08:12:53.4056619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4056684Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4056910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4056984Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4057210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4057282Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4057509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4057683Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4057687Z 2025-09-07T08:12:53.4057776Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4057960Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4058020Z return mod(**inputs) 2025-09-07T08:12:53.4058251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4058315Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4058546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4058607Z layer_outputs = layer_module( 2025-09-07T08:12:53.4058822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4058926Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4059190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4059261Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4059491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4059560Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4059785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4059899Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4059903Z 2025-09-07T08:12:53.4059988Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4060181Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4060238Z return mod(**inputs) 2025-09-07T08:12:53.4060499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4060566Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4060828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4060897Z layer_outputs = layer_module( 2025-09-07T08:12:53.4061109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4061176Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4061408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4061478Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4061712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4061783Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4062017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.4062162Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.4062167Z 2025-09-07T08:12:53.4062258Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4062443Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4062498Z return mod(**inputs) 2025-09-07T08:12:53.4062735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4062799Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4063026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4063096Z layer_outputs = layer_module( 2025-09-07T08:12:53.4063310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4063377Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4063605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4063678Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4063906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4063973Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4064202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.4064336Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.4064340Z 2025-09-07T08:12:53.4064469Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4064682Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4064739Z return mod(**inputs) 2025-09-07T08:12:53.4064971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4065033Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4065276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4065335Z layer_outputs = layer_module( 2025-09-07T08:12:53.4065546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4065613Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4065838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4065909Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4066168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4066242Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4066516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.4066586Z value_states = self.v(current_states) 2025-09-07T08:12:53.4066589Z 2025-09-07T08:12:53.4066682Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4066863Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4066925Z return mod(**inputs) 2025-09-07T08:12:53.4067153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4067216Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4067459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4067525Z layer_outputs = layer_module( 2025-09-07T08:12:53.4067746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4067815Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4068046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4068122Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4068351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4068431Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4068659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4068767Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4068771Z 2025-09-07T08:12:53.4068863Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4069047Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4069113Z return mod(**inputs) 2025-09-07T08:12:53.4069344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4069414Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4069645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4069710Z layer_outputs = layer_module( 2025-09-07T08:12:53.4069928Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4069995Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4070527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4070627Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4070865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4070938Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4071165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4071269Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4071273Z 2025-09-07T08:12:53.4071364Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4071559Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4071617Z return mod(**inputs) 2025-09-07T08:12:53.4071854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4071959Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4072194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4072295Z layer_outputs = layer_module( 2025-09-07T08:12:53.4072511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4072580Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4072815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4072886Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4073121Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4073193Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4073430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4073526Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4073530Z 2025-09-07T08:12:53.4073621Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4073814Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4073873Z return mod(**inputs) 2025-09-07T08:12:53.4074115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4074178Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4074409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4074480Z layer_outputs = layer_module( 2025-09-07T08:12:53.4074692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4074773Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4075001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4075075Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4075309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4075381Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4075617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.4075708Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.4075711Z 2025-09-07T08:12:53.4075806Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4075987Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4076104Z return mod(**inputs) 2025-09-07T08:12:53.4076345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4076410Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4076647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4076710Z layer_outputs = layer_module( 2025-09-07T08:12:53.4076932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4077095Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4077355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4077671Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4077932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4078068Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4078365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.4078493Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.4078496Z 2025-09-07T08:12:53.4078667Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4087096Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4087172Z return mod(**inputs) 2025-09-07T08:12:53.4087447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4087519Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4087770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4087849Z layer_outputs = layer_module( 2025-09-07T08:12:53.4088087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4088169Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4088415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4088496Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4088725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:12:53.4088849Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:12:53.4088853Z 2025-09-07T08:12:53.4088934Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4089035Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4089242Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4089306Z return mod(**inputs) 2025-09-07T08:12:53.4089559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4089636Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4089890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4089961Z layer_outputs = layer_module( 2025-09-07T08:12:53.4090186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4090262Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4090505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4090589Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4090823Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:12:53.4091052Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4091284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4091359Z return self.weight * hidden_states 2025-09-07T08:12:53.4091363Z 2025-09-07T08:12:53.4091463Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4091675Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4091738Z return mod(**inputs) 2025-09-07T08:12:53.4091983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4092051Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4092283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4092356Z layer_outputs = layer_module( 2025-09-07T08:12:53.4092607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4092688Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4092954Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4093038Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4093271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4093384Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4093618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:12:53.4093690Z hidden_states = self.wi(hidden_states) 2025-09-07T08:12:53.4093695Z 2025-09-07T08:12:53.4093795Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4093989Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4094046Z return mod(**inputs) 2025-09-07T08:12:53.4094285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4094351Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4094587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4094649Z layer_outputs = layer_module( 2025-09-07T08:12:53.4094863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4094937Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4095166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4095252Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4095477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4095584Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4095812Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:12:53.4095886Z hidden_states = self.act(hidden_states) 2025-09-07T08:12:53.4095889Z 2025-09-07T08:12:53.4095986Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4096176Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4096237Z return mod(**inputs) 2025-09-07T08:12:53.4096468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4096568Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4096842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4096907Z layer_outputs = layer_module( 2025-09-07T08:12:53.4097129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4097200Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4097428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4097514Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4097812Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4097924Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4098156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:12:53.4098269Z hidden_states = self.wo(hidden_states) 2025-09-07T08:12:53.4098273Z 2025-09-07T08:12:53.4098347Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4098440Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4098662Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4098721Z return mod(**inputs) 2025-09-07T08:12:53.4098956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:12:53.4099020Z encoder_outputs = self.encoder( 2025-09-07T08:12:53.4099249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1128, in forward 2025-09-07T08:12:53.4099346Z hidden_states = self.final_layer_norm(hidden_states) 2025-09-07T08:12:53.4099575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4099650Z return self.weight * hidden_states 2025-09-07T08:12:53.4099654Z 2025-09-07T08:12:53.4099746Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4099934Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4099991Z return mod(**inputs) 2025-09-07T08:12:53.4100218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4100282Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4100511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4100576Z layer_outputs = layer_module( 2025-09-07T08:12:53.4100788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4100859Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4101091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4101161Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4101392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4101465Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4101690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.4101760Z key_states = self.k(current_states) 2025-09-07T08:12:53.4101764Z 2025-09-07T08:12:53.4101852Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4102035Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4102092Z return mod(**inputs) 2025-09-07T08:12:53.4102360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4102450Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4102683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4102746Z layer_outputs = layer_module( 2025-09-07T08:12:53.4102958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4103026Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4103440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4103513Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4103743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4103816Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4104107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4104230Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4104233Z 2025-09-07T08:12:53.4104364Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4104553Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4104612Z return mod(**inputs) 2025-09-07T08:12:53.4104851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4104914Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4105147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4105206Z layer_outputs = layer_module( 2025-09-07T08:12:53.4105422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4105496Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4105726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4105797Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4106034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4106109Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4106335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4106446Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4106450Z 2025-09-07T08:12:53.4106543Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4106729Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4106789Z return mod(**inputs) 2025-09-07T08:12:53.4107026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4107090Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4107327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4107387Z layer_outputs = layer_module( 2025-09-07T08:12:53.4107600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4107665Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4107890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4108008Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4108292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4108367Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4108595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4108704Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4108708Z 2025-09-07T08:12:53.4108797Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4108979Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4109037Z return mod(**inputs) 2025-09-07T08:12:53.4109269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4109332Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4109637Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4109699Z layer_outputs = layer_module( 2025-09-07T08:12:53.4109913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4110023Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4110259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4110329Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4110556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4110632Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4110855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.4111009Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.4111015Z 2025-09-07T08:12:53.4111107Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4111297Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4111358Z return mod(**inputs) 2025-09-07T08:12:53.4111595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4111663Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4111893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4111962Z layer_outputs = layer_module( 2025-09-07T08:12:53.4112174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4112241Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4112479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4112555Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4112794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4112866Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4113093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.4113173Z value_states = self.v(current_states) 2025-09-07T08:12:53.4113177Z 2025-09-07T08:12:53.4113271Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4113464Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4113525Z return mod(**inputs) 2025-09-07T08:12:53.4113758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4113893Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4114124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4114198Z layer_outputs = layer_module( 2025-09-07T08:12:53.4114413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4114488Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4114716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4114785Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4115015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4115085Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4115346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4115449Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4115453Z 2025-09-07T08:12:53.4115543Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4115758Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4115821Z return mod(**inputs) 2025-09-07T08:12:53.4116059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4116121Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4116377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4116441Z layer_outputs = layer_module( 2025-09-07T08:12:53.4116667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4116741Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4116974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4117047Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4117273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4117348Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4117579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4117675Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4117679Z 2025-09-07T08:12:53.4117774Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4117967Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4118031Z return mod(**inputs) 2025-09-07T08:12:53.4118271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4118335Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4118570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4118633Z layer_outputs = layer_module( 2025-09-07T08:12:53.4118850Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4118917Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4119141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4119217Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4119485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4119595Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4119821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4119915Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4119919Z 2025-09-07T08:12:53.4120015Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4120199Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4120262Z return mod(**inputs) 2025-09-07T08:12:53.4120493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4120557Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4120790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4120854Z layer_outputs = layer_module( 2025-09-07T08:12:53.4121102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4121172Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4121437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4121507Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4121736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4121815Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4122043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.4122141Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.4122146Z 2025-09-07T08:12:53.4122236Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4122426Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4122489Z return mod(**inputs) 2025-09-07T08:12:53.4122719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4122786Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4123014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4123080Z layer_outputs = layer_module( 2025-09-07T08:12:53.4123290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4123356Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4123588Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4123662Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4123893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4123967Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4124196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.4124273Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.4124277Z 2025-09-07T08:12:53.4124353Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4124449Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4124633Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4124690Z return mod(**inputs) 2025-09-07T08:12:53.4124924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4125046Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4125283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4125343Z layer_outputs = layer_module( 2025-09-07T08:12:53.4125558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4125625Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4125850Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4125939Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4126165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:12:53.4126258Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4126519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4126590Z return self.weight * hidden_states 2025-09-07T08:12:53.4126593Z 2025-09-07T08:12:53.4126688Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4126903Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4126968Z return mod(**inputs) 2025-09-07T08:12:53.4127200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4127265Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4127499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4127560Z layer_outputs = layer_module( 2025-09-07T08:12:53.4127779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4127847Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4128080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4128159Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4128388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4128501Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4128726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:12:53.4128808Z hidden_states = self.wi(hidden_states) 2025-09-07T08:12:53.4128813Z 2025-09-07T08:12:53.4128903Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4129095Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4129153Z return mod(**inputs) 2025-09-07T08:12:53.4129387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4129448Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4129683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4129744Z layer_outputs = layer_module( 2025-09-07T08:12:53.4129956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4130028Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4130252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4130332Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4130556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4130720Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4130953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:12:53.4131028Z hidden_states = self.act(hidden_states) 2025-09-07T08:12:53.4131031Z 2025-09-07T08:12:53.4131126Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4131308Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4131372Z return mod(**inputs) 2025-09-07T08:12:53.4131604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4131667Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4131900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4131964Z layer_outputs = layer_module( 2025-09-07T08:12:53.4132228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4132296Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4132558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4132643Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4132872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4132977Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4133204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:12:53.4133272Z hidden_states = self.wo(hidden_states) 2025-09-07T08:12:53.4133282Z 2025-09-07T08:12:53.4133370Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4133557Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4133615Z return mod(**inputs) 2025-09-07T08:12:53.4133845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4133913Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4134142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4134201Z layer_outputs = layer_module( 2025-09-07T08:12:53.4134413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4134480Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4134710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4134783Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4135011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:12:53.4135112Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4135341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4135412Z return self.weight * hidden_states 2025-09-07T08:12:53.4135415Z 2025-09-07T08:12:53.4135504Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4135690Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4135747Z return mod(**inputs) 2025-09-07T08:12:53.4135978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4136077Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4136313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4136417Z layer_outputs = layer_module( 2025-09-07T08:12:53.4136632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4136699Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4136933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4137002Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4137234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4137306Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4137532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.4137669Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.4137703Z 2025-09-07T08:12:53.4137795Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4137984Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4138074Z return mod(**inputs) 2025-09-07T08:12:53.4138310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4138378Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4138608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4138676Z layer_outputs = layer_module( 2025-09-07T08:12:53.4138888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4138957Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4139187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4139259Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4139495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4139564Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4139796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.4139866Z key_states = self.k(current_states) 2025-09-07T08:12:53.4139869Z 2025-09-07T08:12:53.4139959Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4140145Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4140201Z return mod(**inputs) 2025-09-07T08:12:53.4140434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4140500Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4140737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4140798Z layer_outputs = layer_module( 2025-09-07T08:12:53.4141016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4141085Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4141310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4141380Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4141605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4141674Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4141940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4142087Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4142091Z 2025-09-07T08:12:53.4142184Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4142368Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4142423Z return mod(**inputs) 2025-09-07T08:12:53.4142659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4142720Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4142955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4143014Z layer_outputs = layer_module( 2025-09-07T08:12:53.4143227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4143334Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4143566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4143671Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4143901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4143979Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4144206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4144320Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4144323Z 2025-09-07T08:12:53.4144419Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4144605Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4144670Z return mod(**inputs) 2025-09-07T08:12:53.4144903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4144966Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4145200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4145262Z layer_outputs = layer_module( 2025-09-07T08:12:53.4145481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4145548Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4145780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4145850Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4146082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4146160Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4146384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4146499Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4146502Z 2025-09-07T08:12:53.4146593Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4146773Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4146834Z return mod(**inputs) 2025-09-07T08:12:53.4147061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4147129Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4147359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4147494Z layer_outputs = layer_module( 2025-09-07T08:12:53.4147707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4147773Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4148005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4148075Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4148307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4148378Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4148605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.4148756Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.4148761Z 2025-09-07T08:12:53.4148882Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4149069Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4149126Z return mod(**inputs) 2025-09-07T08:12:53.4149550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4149621Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4149849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4149915Z layer_outputs = layer_module( 2025-09-07T08:12:53.4150124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4150195Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4150422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4150496Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4150725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4150796Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4151026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.4151096Z value_states = self.v(current_states) 2025-09-07T08:12:53.4151099Z 2025-09-07T08:12:53.4151187Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4151377Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4151434Z return mod(**inputs) 2025-09-07T08:12:53.4151668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4151731Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4151963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4152024Z layer_outputs = layer_module( 2025-09-07T08:12:53.4152235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4152309Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4152536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4152607Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4152836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4152907Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4153139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4153314Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4153318Z 2025-09-07T08:12:53.4153412Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4153594Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4153652Z return mod(**inputs) 2025-09-07T08:12:53.4153888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4153949Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4154180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4154241Z layer_outputs = layer_module( 2025-09-07T08:12:53.4154457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4154527Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4154788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4154869Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4155127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4155200Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4155428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4155519Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4155522Z 2025-09-07T08:12:53.4155616Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4155803Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4155866Z return mod(**inputs) 2025-09-07T08:12:53.4156096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4156159Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4156390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4156451Z layer_outputs = layer_module( 2025-09-07T08:12:53.4156669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4156736Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4156969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4157038Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4157263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4157341Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4157570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4157663Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4157667Z 2025-09-07T08:12:53.4157755Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4157937Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4158001Z return mod(**inputs) 2025-09-07T08:12:53.4158229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4158296Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4158522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4158619Z layer_outputs = layer_module( 2025-09-07T08:12:53.4158833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4158932Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4159168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4159237Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4159469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4159537Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4159761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.4159857Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.4159862Z 2025-09-07T08:12:53.4159952Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4160168Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4160228Z return mod(**inputs) 2025-09-07T08:12:53.4160461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4160556Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4160793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4160860Z layer_outputs = layer_module( 2025-09-07T08:12:53.4161074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4161148Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4161376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4161447Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4161679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4161750Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4161982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.4162051Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.4162054Z 2025-09-07T08:12:53.4162126Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4162221Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4162406Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4162467Z return mod(**inputs) 2025-09-07T08:12:53.4162699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4162761Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4162998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4163059Z layer_outputs = layer_module( 2025-09-07T08:12:53.4163274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4163340Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4163573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4163644Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4163870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T08:12:53.4163973Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4164200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4164326Z return self.weight * hidden_states 2025-09-07T08:12:53.4164331Z 2025-09-07T08:12:53.4164419Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4164602Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4164668Z return mod(**inputs) 2025-09-07T08:12:53.4164898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4164965Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4165195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4165258Z layer_outputs = layer_module( 2025-09-07T08:12:53.4165475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4165541Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4165807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4165879Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4166142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4166217Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4166444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.4166518Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.4166521Z 2025-09-07T08:12:53.4166610Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4166794Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4166850Z return mod(**inputs) 2025-09-07T08:12:53.4167082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4167157Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4167384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4167450Z layer_outputs = layer_module( 2025-09-07T08:12:53.4167660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4167726Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4167955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4168025Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4168252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4168324Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4168553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.4168621Z key_states = self.k(current_states) 2025-09-07T08:12:53.4168624Z 2025-09-07T08:12:53.4168713Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4168900Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4168957Z return mod(**inputs) 2025-09-07T08:12:53.4169189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4169252Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4169481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4169545Z layer_outputs = layer_module( 2025-09-07T08:12:53.4169757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4169889Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4170115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4170186Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4170414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4170486Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4170713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4170826Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4170830Z 2025-09-07T08:12:53.4170920Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4171099Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4171157Z return mod(**inputs) 2025-09-07T08:12:53.4171423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4171486Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4171756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4171817Z layer_outputs = layer_module( 2025-09-07T08:12:53.4172028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4172100Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4172329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4172402Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4172630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4172713Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4172938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4173050Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4173053Z 2025-09-07T08:12:53.4173147Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4173328Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4173386Z return mod(**inputs) 2025-09-07T08:12:53.4173615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4173677Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4173909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4173974Z layer_outputs = layer_module( 2025-09-07T08:12:53.4174190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4174254Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4174481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4174556Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4174779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4174857Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4175081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4175193Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4175242Z 2025-09-07T08:12:53.4175331Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4175546Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4175605Z return mod(**inputs) 2025-09-07T08:12:53.4175838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4175905Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4176132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4176191Z layer_outputs = layer_module( 2025-09-07T08:12:53.4176408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4176473Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4176701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4176804Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4177040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4177111Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4177367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.4177518Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.4177522Z 2025-09-07T08:12:53.4177645Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4177832Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4177887Z return mod(**inputs) 2025-09-07T08:12:53.4178116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4178190Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4178421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4178486Z layer_outputs = layer_module( 2025-09-07T08:12:53.4178701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4178767Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4178999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4179066Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4179296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4179367Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4179596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.4179669Z value_states = self.v(current_states) 2025-09-07T08:12:53.4179672Z 2025-09-07T08:12:53.4179759Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4179948Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4180003Z return mod(**inputs) 2025-09-07T08:12:53.4180237Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4180300Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4180529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4180592Z layer_outputs = layer_module( 2025-09-07T08:12:53.4180804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4180914Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4181180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4181253Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4181485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4181556Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4181788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4181883Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4181886Z 2025-09-07T08:12:53.4181987Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4182168Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4182226Z return mod(**inputs) 2025-09-07T08:12:53.4182498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4182563Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4182822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4182885Z layer_outputs = layer_module( 2025-09-07T08:12:53.4183098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4183171Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4183397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4183471Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4183698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4183778Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4184007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4184098Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4184102Z 2025-09-07T08:12:53.4184198Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4184382Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4184443Z return mod(**inputs) 2025-09-07T08:12:53.4184670Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4184733Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4184969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4185034Z layer_outputs = layer_module( 2025-09-07T08:12:53.4185250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4185317Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4185547Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4185622Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4185848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4185923Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4186149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4186245Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4186247Z 2025-09-07T08:12:53.4186334Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4186550Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4186642Z return mod(**inputs) 2025-09-07T08:12:53.4186874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4186942Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4187169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4187231Z layer_outputs = layer_module( 2025-09-07T08:12:53.4187450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4187519Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4187747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4187818Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4188079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4188156Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4188416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.4188516Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.4188520Z 2025-09-07T08:12:53.4188610Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4188798Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4188855Z return mod(**inputs) 2025-09-07T08:12:53.4189086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4189153Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4189384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4189451Z layer_outputs = layer_module( 2025-09-07T08:12:53.4189665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4189736Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4189963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4190034Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4190268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4190340Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4190566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.4190640Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.4190645Z 2025-09-07T08:12:53.4190717Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4190808Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4190992Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4191054Z return mod(**inputs) 2025-09-07T08:12:53.4191283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4191344Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4191580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4191640Z layer_outputs = layer_module( 2025-09-07T08:12:53.4191858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4191962Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4192223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4192312Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4192540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:12:53.4192633Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4192856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4192928Z return self.weight * hidden_states 2025-09-07T08:12:53.4192932Z 2025-09-07T08:12:53.4193023Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4193212Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4193276Z return mod(**inputs) 2025-09-07T08:12:53.4193538Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4193606Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4193838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4193948Z layer_outputs = layer_module( 2025-09-07T08:12:53.4194170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4194238Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4194471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4194547Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4194772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4194887Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4195117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:12:53.4195191Z hidden_states = self.wi(hidden_states) 2025-09-07T08:12:53.4195194Z 2025-09-07T08:12:53.4195288Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4195478Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4195535Z return mod(**inputs) 2025-09-07T08:12:53.4195792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4195874Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4196103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4196167Z layer_outputs = layer_module( 2025-09-07T08:12:53.4196382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4196449Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4196683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4196763Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4196992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4197093Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4197325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:12:53.4197402Z hidden_states = self.act(hidden_states) 2025-09-07T08:12:53.4197405Z 2025-09-07T08:12:53.4197493Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4197719Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4197848Z return mod(**inputs) 2025-09-07T08:12:53.4198083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4198148Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4198381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4198447Z layer_outputs = layer_module( 2025-09-07T08:12:53.4198660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4198730Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4198958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4199036Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4199299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4199399Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4199665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:12:53.4199738Z hidden_states = self.wo(hidden_states) 2025-09-07T08:12:53.4199741Z 2025-09-07T08:12:53.4199815Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4199902Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4200083Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4200145Z return mod(**inputs) 2025-09-07T08:12:53.4200377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4200445Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4200676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4200739Z layer_outputs = layer_module( 2025-09-07T08:12:53.4200960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4201026Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4201255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4201330Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4201559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:12:53.4201660Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4201884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4201960Z return self.weight * hidden_states 2025-09-07T08:12:53.4201964Z 2025-09-07T08:12:53.4202052Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4202239Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4202296Z return mod(**inputs) 2025-09-07T08:12:53.4202523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4202591Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4202818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4202879Z layer_outputs = layer_module( 2025-09-07T08:12:53.4203218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4203288Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4203572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4203689Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4203921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4203996Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4204223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.4204295Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.4204299Z 2025-09-07T08:12:53.4204388Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4204574Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4204634Z return mod(**inputs) 2025-09-07T08:12:53.4204867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4204971Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4205207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4205313Z layer_outputs = layer_module( 2025-09-07T08:12:53.4205533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4205604Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4205840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4205914Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4206146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4206222Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4206459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.4206526Z key_states = self.k(current_states) 2025-09-07T08:12:53.4206530Z 2025-09-07T08:12:53.4206632Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4206825Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4206883Z return mod(**inputs) 2025-09-07T08:12:53.4207127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4207193Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4207429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4207493Z layer_outputs = layer_module( 2025-09-07T08:12:53.4207710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4207789Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4208018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4208096Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4208326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4208400Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4208632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4208755Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4208758Z 2025-09-07T08:12:53.4208858Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4209045Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4209170Z return mod(**inputs) 2025-09-07T08:12:53.4209407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4209471Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4209710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4209772Z layer_outputs = layer_module( 2025-09-07T08:12:53.4209996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4210064Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4210295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4210375Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4210604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4210717Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4210947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4211088Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4211098Z 2025-09-07T08:12:53.4211190Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4211376Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4211440Z return mod(**inputs) 2025-09-07T08:12:53.4211671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4211741Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4211979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4212047Z layer_outputs = layer_module( 2025-09-07T08:12:53.4212269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4212337Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4212570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4212639Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4212867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4212946Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4213174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4213290Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4213295Z 2025-09-07T08:12:53.4213384Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4213577Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4213635Z return mod(**inputs) 2025-09-07T08:12:53.4213866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4213936Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4214166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4214234Z layer_outputs = layer_module( 2025-09-07T08:12:53.4214445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4214512Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4214745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4214875Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4215110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4215180Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4215407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.4215562Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.4215566Z 2025-09-07T08:12:53.4215653Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4215843Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4215900Z return mod(**inputs) 2025-09-07T08:12:53.4216135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4216201Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4216480Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4216550Z layer_outputs = layer_module( 2025-09-07T08:12:53.4216794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4216869Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4217099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4217170Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4217406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4217476Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4217763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.4217839Z value_states = self.v(current_states) 2025-09-07T08:12:53.4217842Z 2025-09-07T08:12:53.4217938Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4218122Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4218179Z return mod(**inputs) 2025-09-07T08:12:53.4218429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4218491Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4218728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4218793Z layer_outputs = layer_module( 2025-09-07T08:12:53.4219008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4219087Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4219320Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4219396Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4219626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4219707Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4219945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4220046Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4220049Z 2025-09-07T08:12:53.4220147Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4220331Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4220429Z return mod(**inputs) 2025-09-07T08:12:53.4220665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4220760Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4221001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4221063Z layer_outputs = layer_module( 2025-09-07T08:12:53.4221284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4221353Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4221583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4221660Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4221889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4221969Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4222232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4222330Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4222339Z 2025-09-07T08:12:53.4222457Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4222643Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4222711Z return mod(**inputs) 2025-09-07T08:12:53.4222945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4223018Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4223254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4223314Z layer_outputs = layer_module( 2025-09-07T08:12:53.4223539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4223607Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4223841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4223910Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4224137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4224211Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4224438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4224535Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4224538Z 2025-09-07T08:12:53.4224624Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4224812Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4224871Z return mod(**inputs) 2025-09-07T08:12:53.4225101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4225170Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4225416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4225483Z layer_outputs = layer_module( 2025-09-07T08:12:53.4225696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4225763Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4225992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4226062Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4226331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4226431Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4226663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.4226758Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.4226762Z 2025-09-07T08:12:53.4226848Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4227036Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4227093Z return mod(**inputs) 2025-09-07T08:12:53.4227327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4227387Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4227620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4227717Z layer_outputs = layer_module( 2025-09-07T08:12:53.4227933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4228035Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4228268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4228339Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4228570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4228640Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4228871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.4228940Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.4228943Z 2025-09-07T08:12:53.4229033Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4229220Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4229277Z return mod(**inputs) 2025-09-07T08:12:53.4229508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4229572Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4229804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4229864Z layer_outputs = layer_module( 2025-09-07T08:12:53.4230073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4230141Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4230365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4230440Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4230666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:12:53.4230789Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:12:53.4230793Z 2025-09-07T08:12:53.4230866Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4230954Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4231140Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4231195Z return mod(**inputs) 2025-09-07T08:12:53.4231425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4231489Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4231752Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4231849Z layer_outputs = layer_module( 2025-09-07T08:12:53.4232064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4232134Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4232359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4232430Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4232658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T08:12:53.4232755Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4232985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4233054Z return self.weight * hidden_states 2025-09-07T08:12:53.4233058Z 2025-09-07T08:12:53.4233178Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4233364Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4233419Z return mod(**inputs) 2025-09-07T08:12:53.4233681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4233746Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4233974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4234038Z layer_outputs = layer_module( 2025-09-07T08:12:53.4234246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4234315Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4234539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4234617Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4234843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4234915Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4235145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.4235212Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.4235215Z 2025-09-07T08:12:53.4235313Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4235494Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4235548Z return mod(**inputs) 2025-09-07T08:12:53.4235778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4235843Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4236074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4236133Z layer_outputs = layer_module( 2025-09-07T08:12:53.4236347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4236415Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4236643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4236720Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4236945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4237021Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4237247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.4237392Z key_states = self.k(current_states) 2025-09-07T08:12:53.4237395Z 2025-09-07T08:12:53.4237486Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4237667Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4237725Z return mod(**inputs) 2025-09-07T08:12:53.4237953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4238015Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4238247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4238308Z layer_outputs = layer_module( 2025-09-07T08:12:53.4238523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4238592Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4238854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4238927Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4239182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4239261Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4239490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4239612Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4239615Z 2025-09-07T08:12:53.4239705Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4239887Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4239951Z return mod(**inputs) 2025-09-07T08:12:53.4240185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4240253Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4240483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4240545Z layer_outputs = layer_module( 2025-09-07T08:12:53.4240757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4240826Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4241055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4241122Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4241351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4241425Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4241653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4241765Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4241770Z 2025-09-07T08:12:53.4241859Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4242045Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4242102Z return mod(**inputs) 2025-09-07T08:12:53.4242333Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4242403Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4242636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4242735Z layer_outputs = layer_module( 2025-09-07T08:12:53.4242979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4243048Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4243279Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4243348Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4243575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4243646Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4243879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4243992Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4243996Z 2025-09-07T08:12:53.4244088Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4244311Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4244370Z return mod(**inputs) 2025-09-07T08:12:53.4244638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4244703Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4244938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4245007Z layer_outputs = layer_module( 2025-09-07T08:12:53.4245219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4245290Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4245515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4245592Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4245823Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4245897Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4246127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.4246274Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.4246278Z 2025-09-07T08:12:53.4246371Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4246553Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4246611Z return mod(**inputs) 2025-09-07T08:12:53.4246844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4246906Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4247140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4247199Z layer_outputs = layer_module( 2025-09-07T08:12:53.4247408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4247477Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4247702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4247783Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4248010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4248089Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4248314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.4248448Z value_states = self.v(current_states) 2025-09-07T08:12:53.4248452Z 2025-09-07T08:12:53.4248549Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4248731Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4248799Z return mod(**inputs) 2025-09-07T08:12:53.4249031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4249092Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4249328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4249391Z layer_outputs = layer_module( 2025-09-07T08:12:53.4249608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4249674Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4249935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4250013Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4250270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4250351Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4250578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4250679Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4250682Z 2025-09-07T08:12:53.4250771Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4250954Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4251019Z return mod(**inputs) 2025-09-07T08:12:53.4251254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4251324Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4251555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4251622Z layer_outputs = layer_module( 2025-09-07T08:12:53.4251842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4251909Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4252142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4252211Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4252439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4252513Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4252740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4252834Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4252837Z 2025-09-07T08:12:53.4252924Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4253112Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4253168Z return mod(**inputs) 2025-09-07T08:12:53.4253398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4253468Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4253700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4253768Z layer_outputs = layer_module( 2025-09-07T08:12:53.4254020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4254120Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4254364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4254440Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4254675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4254749Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4254987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4255081Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4255086Z 2025-09-07T08:12:53.4255176Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4255369Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4255458Z return mod(**inputs) 2025-09-07T08:12:53.4255695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4255759Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4256021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4256092Z layer_outputs = layer_module( 2025-09-07T08:12:53.4256318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4256394Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4256643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4256714Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4256952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4257027Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4257260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.4257354Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.4257359Z 2025-09-07T08:12:53.4257454Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4257695Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4257754Z return mod(**inputs) 2025-09-07T08:12:53.4257994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4258057Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4258290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4258357Z layer_outputs = layer_module( 2025-09-07T08:12:53.4258568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4258640Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4258871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4258948Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4259176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4259258Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4259486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.4259554Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.4259606Z 2025-09-07T08:12:53.4259681Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4259800Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4259987Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4260041Z return mod(**inputs) 2025-09-07T08:12:53.4260271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4260339Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4260567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4260636Z layer_outputs = layer_module( 2025-09-07T08:12:53.4260849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4260915Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4261185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4261271Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4261509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:12:53.4261629Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4261866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4261936Z return self.weight * hidden_states 2025-09-07T08:12:53.4261939Z 2025-09-07T08:12:53.4262032Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4262224Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4262282Z return mod(**inputs) 2025-09-07T08:12:53.4262520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4262588Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4262817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4262882Z layer_outputs = layer_module( 2025-09-07T08:12:53.4263095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4263167Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4263392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4263468Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4263696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4263802Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4264039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:12:53.4264111Z hidden_states = self.wi(hidden_states) 2025-09-07T08:12:53.4264114Z 2025-09-07T08:12:53.4264203Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4264385Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4264440Z return mod(**inputs) 2025-09-07T08:12:53.4264670Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4264731Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4264961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4265020Z layer_outputs = layer_module( 2025-09-07T08:12:53.4265236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4265372Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4265601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4265687Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4265915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4266018Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4266254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:12:53.4266328Z hidden_states = self.act(hidden_states) 2025-09-07T08:12:53.4266331Z 2025-09-07T08:12:53.4266429Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4266611Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4266676Z return mod(**inputs) 2025-09-07T08:12:53.4266936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4266998Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4267261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4267321Z layer_outputs = layer_module( 2025-09-07T08:12:53.4267537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4267603Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4267832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4267910Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4268134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4268240Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4268464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:12:53.4268540Z hidden_states = self.wo(hidden_states) 2025-09-07T08:12:53.4268544Z 2025-09-07T08:12:53.4268615Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4268706Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4268898Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4268954Z return mod(**inputs) 2025-09-07T08:12:53.4269191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4269254Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4269488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4269563Z layer_outputs = layer_module( 2025-09-07T08:12:53.4269776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4269852Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4270079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4270154Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4270387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:12:53.4270483Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4270719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4270821Z return self.weight * hidden_states 2025-09-07T08:12:53.4270851Z 2025-09-07T08:12:53.4270944Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4271129Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4271184Z return mod(**inputs) 2025-09-07T08:12:53.4271418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4271478Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4271719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4271785Z layer_outputs = layer_module( 2025-09-07T08:12:53.4271994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4272065Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4272295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4272405Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4272642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4272752Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4273001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.4273070Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.4273074Z 2025-09-07T08:12:53.4273174Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4273356Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4273429Z return mod(**inputs) 2025-09-07T08:12:53.4273661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4273729Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4273969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4274031Z layer_outputs = layer_module( 2025-09-07T08:12:53.4274251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4274325Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4274561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4274638Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4274865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4274938Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4275160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.4275234Z key_states = self.k(current_states) 2025-09-07T08:12:53.4275241Z 2025-09-07T08:12:53.4275332Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4275516Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4275573Z return mod(**inputs) 2025-09-07T08:12:53.4275800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4275864Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4276092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4276151Z layer_outputs = layer_module( 2025-09-07T08:12:53.4276361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4276459Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4276717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4276785Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4277012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4277085Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4277310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4277431Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4277435Z 2025-09-07T08:12:53.4277521Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4277705Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4277765Z return mod(**inputs) 2025-09-07T08:12:53.4278027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4278092Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4278369Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4278432Z layer_outputs = layer_module( 2025-09-07T08:12:53.4278646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4278716Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4278954Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4279025Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4279260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4279336Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4279567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4279685Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4279689Z 2025-09-07T08:12:53.4279779Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4279969Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4280027Z return mod(**inputs) 2025-09-07T08:12:53.4280265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4280326Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4280553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4280617Z layer_outputs = layer_module( 2025-09-07T08:12:53.4280830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4280896Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4281125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4281194Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4281421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4281491Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4281720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4281828Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4281832Z 2025-09-07T08:12:53.4281957Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4282166Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4282221Z return mod(**inputs) 2025-09-07T08:12:53.4282453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4282515Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4282743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4282802Z layer_outputs = layer_module( 2025-09-07T08:12:53.4283013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4283080Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4283305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4283377Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4283645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4283721Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4283988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.4284138Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.4284142Z 2025-09-07T08:12:53.4284241Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4284434Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4284493Z return mod(**inputs) 2025-09-07T08:12:53.4284727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4284793Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4285027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4285091Z layer_outputs = layer_module( 2025-09-07T08:12:53.4285315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4285386Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4285614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4285688Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4285914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4285985Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4286209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.4286278Z value_states = self.v(current_states) 2025-09-07T08:12:53.4286286Z 2025-09-07T08:12:53.4286377Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4286563Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4286623Z return mod(**inputs) 2025-09-07T08:12:53.4286852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4286916Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4287149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4287209Z layer_outputs = layer_module( 2025-09-07T08:12:53.4287423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4287491Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4287803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4287903Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4288130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4288205Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4288429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4288535Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4288539Z 2025-09-07T08:12:53.4288629Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4288819Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4288877Z return mod(**inputs) 2025-09-07T08:12:53.4289109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4289208Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4289444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4289543Z layer_outputs = layer_module( 2025-09-07T08:12:53.4289759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4289828Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4290065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4290136Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4290370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4290442Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4290674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4290778Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4290782Z 2025-09-07T08:12:53.4290873Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4291064Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4291121Z return mod(**inputs) 2025-09-07T08:12:53.4291360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4291426Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4291657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4291726Z layer_outputs = layer_module( 2025-09-07T08:12:53.4291940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4292019Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4292247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4292318Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4292550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4292622Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4292857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4292946Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4292951Z 2025-09-07T08:12:53.4293040Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4293221Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4293344Z return mod(**inputs) 2025-09-07T08:12:53.4293576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4293637Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4293868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4293928Z layer_outputs = layer_module( 2025-09-07T08:12:53.4294138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4294205Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4294430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4294500Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4294726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4294826Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4295057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.4295179Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.4295182Z 2025-09-07T08:12:53.4295272Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4295452Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4295510Z return mod(**inputs) 2025-09-07T08:12:53.4295739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4295799Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4296029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4296092Z layer_outputs = layer_module( 2025-09-07T08:12:53.4296305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4296369Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4296595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4296665Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4296890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4296959Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4297182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.4297247Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.4297254Z 2025-09-07T08:12:53.4297325Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4297414Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4297667Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4297722Z return mod(**inputs) 2025-09-07T08:12:53.4297957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4298018Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4298246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4298308Z layer_outputs = layer_module( 2025-09-07T08:12:53.4298518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4298584Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4298808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4298942Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4299171Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T08:12:53.4299269Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4299506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4299575Z return self.weight * hidden_states 2025-09-07T08:12:53.4299578Z 2025-09-07T08:12:53.4299666Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4299858Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4299914Z return mod(**inputs) 2025-09-07T08:12:53.4300154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4307286Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4307707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4307782Z layer_outputs = layer_module( 2025-09-07T08:12:53.4308088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4308165Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4308418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4308493Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4308744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4308827Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4309074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.4309154Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.4309159Z 2025-09-07T08:12:53.4309266Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4309475Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4309536Z return mod(**inputs) 2025-09-07T08:12:53.4309796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4309869Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4310099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4310169Z layer_outputs = layer_module( 2025-09-07T08:12:53.4310390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4310471Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4310711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4310786Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4311018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4311091Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4311319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.4311389Z key_states = self.k(current_states) 2025-09-07T08:12:53.4311393Z 2025-09-07T08:12:53.4311489Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4311690Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4311798Z return mod(**inputs) 2025-09-07T08:12:53.4312041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4312155Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4312390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4312461Z layer_outputs = layer_module( 2025-09-07T08:12:53.4312678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4312753Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4312983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4313056Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4313287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4313364Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4313631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4313759Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4313790Z 2025-09-07T08:12:53.4313893Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4314091Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4314154Z return mod(**inputs) 2025-09-07T08:12:53.4314404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4314469Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4314706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4314771Z layer_outputs = layer_module( 2025-09-07T08:12:53.4314994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4315073Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4315311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4315386Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4315612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4315687Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4315916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4316031Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4316035Z 2025-09-07T08:12:53.4316129Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4316317Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4316381Z return mod(**inputs) 2025-09-07T08:12:53.4316613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4316685Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4316917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4316979Z layer_outputs = layer_module( 2025-09-07T08:12:53.4317198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4317265Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4317491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4317603Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4317862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4317938Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4318169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4318282Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4318286Z 2025-09-07T08:12:53.4318376Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4318566Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4318624Z return mod(**inputs) 2025-09-07T08:12:53.4318852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4318915Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4319176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4319238Z layer_outputs = layer_module( 2025-09-07T08:12:53.4319453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4319551Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4319784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4319853Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4320079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4320151Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4320382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.4320545Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.4320551Z 2025-09-07T08:12:53.4320642Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4320836Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4320893Z return mod(**inputs) 2025-09-07T08:12:53.4321129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4321194Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4321425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4321487Z layer_outputs = layer_module( 2025-09-07T08:12:53.4321704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4321773Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4322007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4322077Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4322311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4322382Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4322611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.4322680Z value_states = self.v(current_states) 2025-09-07T08:12:53.4322684Z 2025-09-07T08:12:53.4322772Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4322959Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4323014Z return mod(**inputs) 2025-09-07T08:12:53.4323284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4323395Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4323627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4323691Z layer_outputs = layer_module( 2025-09-07T08:12:53.4323903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4323973Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4324196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4324265Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4324491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4324565Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4324830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4324936Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4324940Z 2025-09-07T08:12:53.4325072Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4325266Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4325334Z return mod(**inputs) 2025-09-07T08:12:53.4325582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4325652Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4325885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4325947Z layer_outputs = layer_module( 2025-09-07T08:12:53.4326164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4326245Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4326479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4326556Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4326784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4326856Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4327087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4327180Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4327184Z 2025-09-07T08:12:53.4327278Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4327460Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4327523Z return mod(**inputs) 2025-09-07T08:12:53.4327754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4327818Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4328054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4328116Z layer_outputs = layer_module( 2025-09-07T08:12:53.4328335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4328403Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4328629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4328703Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4328968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4329097Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4329324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4329426Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4329429Z 2025-09-07T08:12:53.4329525Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4329710Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4329772Z return mod(**inputs) 2025-09-07T08:12:53.4330002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4330072Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4330302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4330400Z layer_outputs = layer_module( 2025-09-07T08:12:53.4330622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4330691Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4330958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4331029Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4331258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4331330Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4331556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.4331651Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.4331658Z 2025-09-07T08:12:53.4331748Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4331931Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4331986Z return mod(**inputs) 2025-09-07T08:12:53.4332216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4332279Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4332507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4332571Z layer_outputs = layer_module( 2025-09-07T08:12:53.4332781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4332845Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4333073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4333146Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4333372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4333443Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4333666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.4333737Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.4333740Z 2025-09-07T08:12:53.4333827Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4334009Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4334063Z return mod(**inputs) 2025-09-07T08:12:53.4334292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4334391Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4334651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4334717Z layer_outputs = layer_module( 2025-09-07T08:12:53.4334931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4335000Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4335228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4335301Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4335533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 647, in forward 2025-09-07T08:12:53.4335656Z layer_output = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:12:53.4335660Z 2025-09-07T08:12:53.4335743Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4335871Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4336065Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4336126Z return mod(**inputs) 2025-09-07T08:12:53.4336389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4336463Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4336695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4336765Z layer_outputs = layer_module( 2025-09-07T08:12:53.4336980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4337049Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4337284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4337371Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4337694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:12:53.4337784Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4338015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4338092Z return self.weight * hidden_states 2025-09-07T08:12:53.4338096Z 2025-09-07T08:12:53.4338187Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4338380Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4338438Z return mod(**inputs) 2025-09-07T08:12:53.4338678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4338744Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4338978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4339049Z layer_outputs = layer_module( 2025-09-07T08:12:53.4339265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4339338Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4339566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4339646Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4339880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4339994Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4340268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:12:53.4340377Z hidden_states = self.wi(hidden_states) 2025-09-07T08:12:53.4340380Z 2025-09-07T08:12:53.4340471Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4340661Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4340719Z return mod(**inputs) 2025-09-07T08:12:53.4340957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4341021Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4341257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4341319Z layer_outputs = layer_module( 2025-09-07T08:12:53.4341532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4341609Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4341868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4341949Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4342207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4342311Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4342545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:12:53.4342617Z hidden_states = self.act(hidden_states) 2025-09-07T08:12:53.4342621Z 2025-09-07T08:12:53.4342716Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4342902Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4342968Z return mod(**inputs) 2025-09-07T08:12:53.4343202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4343267Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4343507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4343570Z layer_outputs = layer_module( 2025-09-07T08:12:53.4343786Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4343853Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4344081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4344159Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4344387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4344493Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4344717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:12:53.4344785Z hidden_states = self.wo(hidden_states) 2025-09-07T08:12:53.4344793Z 2025-09-07T08:12:53.4344865Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4344952Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4345136Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4345191Z return mod(**inputs) 2025-09-07T08:12:53.4345426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4345487Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4345715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4345846Z layer_outputs = layer_module( 2025-09-07T08:12:53.4346060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4346132Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4346364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4346435Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4346663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:12:53.4346761Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4346988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4347055Z return self.weight * hidden_states 2025-09-07T08:12:53.4347061Z 2025-09-07T08:12:53.4347148Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4347368Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4347425Z return mod(**inputs) 2025-09-07T08:12:53.4347688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4347750Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4347983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4348046Z layer_outputs = layer_module( 2025-09-07T08:12:53.4348257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4348326Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4348552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4348629Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4348854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4348925Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4349153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.4349228Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.4349231Z 2025-09-07T08:12:53.4349322Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4349511Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4349569Z return mod(**inputs) 2025-09-07T08:12:53.4349798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4349863Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4350093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4350159Z layer_outputs = layer_module( 2025-09-07T08:12:53.4350373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4350442Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4350674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4350745Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4350977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4351048Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4351276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.4351435Z key_states = self.k(current_states) 2025-09-07T08:12:53.4351440Z 2025-09-07T08:12:53.4351529Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4351716Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4351772Z return mod(**inputs) 2025-09-07T08:12:53.4352006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4352067Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4352297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4352365Z layer_outputs = layer_module( 2025-09-07T08:12:53.4352577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4352649Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4352909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4352984Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4353248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4353321Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4353551Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4353672Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4353676Z 2025-09-07T08:12:53.4353769Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4353953Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4354010Z return mod(**inputs) 2025-09-07T08:12:53.4354247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4354316Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4354549Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4354611Z layer_outputs = layer_module( 2025-09-07T08:12:53.4354825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4354895Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4355123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4355194Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4355419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4355490Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4355721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4355834Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4355837Z 2025-09-07T08:12:53.4355933Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4356116Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4356175Z return mod(**inputs) 2025-09-07T08:12:53.4356411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4356478Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4356708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4356767Z layer_outputs = layer_module( 2025-09-07T08:12:53.4357018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4357111Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4357337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4357408Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4357632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4357703Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4357930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4358036Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4358042Z 2025-09-07T08:12:53.4358128Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4358309Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4358399Z return mod(**inputs) 2025-09-07T08:12:53.4358634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4358730Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4358964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4359025Z layer_outputs = layer_module( 2025-09-07T08:12:53.4359239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4359306Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4359552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4359625Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4359853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4359930Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4360156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.4360305Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.4360309Z 2025-09-07T08:12:53.4360398Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4360582Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4360637Z return mod(**inputs) 2025-09-07T08:12:53.4360867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4360933Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4361167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4361233Z layer_outputs = layer_module( 2025-09-07T08:12:53.4361442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4361508Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4361735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4361803Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4362030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4362098Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4362322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.4362426Z value_states = self.v(current_states) 2025-09-07T08:12:53.4362458Z 2025-09-07T08:12:53.4362547Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4362730Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4362784Z return mod(**inputs) 2025-09-07T08:12:53.4363017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4363079Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4363308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4363370Z layer_outputs = layer_module( 2025-09-07T08:12:53.4363579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4363646Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4363870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4363971Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4364202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4364298Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4364532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4364634Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4364637Z 2025-09-07T08:12:53.4364726Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4364915Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4364971Z return mod(**inputs) 2025-09-07T08:12:53.4365208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4365272Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4365503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4365563Z layer_outputs = layer_module( 2025-09-07T08:12:53.4365775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4365842Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4366070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4366143Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4366373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4366443Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4366676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4366772Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4366775Z 2025-09-07T08:12:53.4366863Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4367046Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4367102Z return mod(**inputs) 2025-09-07T08:12:53.4367331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4367392Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4367622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4367680Z layer_outputs = layer_module( 2025-09-07T08:12:53.4367891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4368014Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4368254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4368324Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4368552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4368623Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4368847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4368938Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4368943Z 2025-09-07T08:12:53.4369033Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4369219Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4369280Z return mod(**inputs) 2025-09-07T08:12:53.4369545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4369614Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4369889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4369950Z layer_outputs = layer_module( 2025-09-07T08:12:53.4370163Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4370228Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4370457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4370525Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4370750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4370826Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4371050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.4371146Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.4371151Z 2025-09-07T08:12:53.4371239Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4371424Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4371481Z return mod(**inputs) 2025-09-07T08:12:53.4371708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4371770Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4371998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4372060Z layer_outputs = layer_module( 2025-09-07T08:12:53.4372273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4372338Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4372570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4372644Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4372880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4372948Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4373171Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.4373240Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.4373244Z 2025-09-07T08:12:53.4373314Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4373644Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4373859Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4373915Z return mod(**inputs) 2025-09-07T08:12:53.4374148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4374209Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4374444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4374504Z layer_outputs = layer_module( 2025-09-07T08:12:53.4374716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4374780Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4375005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4375082Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4375341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T08:12:53.4375442Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4375699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4375768Z return self.weight * hidden_states 2025-09-07T08:12:53.4375771Z 2025-09-07T08:12:53.4375861Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4376041Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4376097Z return mod(**inputs) 2025-09-07T08:12:53.4376327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4376392Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4376625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4376687Z layer_outputs = layer_module( 2025-09-07T08:12:53.4376900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4376964Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4377193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4377263Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4377489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4377633Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4377860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.4377938Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.4377942Z 2025-09-07T08:12:53.4378028Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4378211Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4378277Z return mod(**inputs) 2025-09-07T08:12:53.4378504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4378567Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4378793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4378855Z layer_outputs = layer_module( 2025-09-07T08:12:53.4379084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4379195Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4379459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4379529Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4379760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4379830Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4380052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.4380120Z key_states = self.k(current_states) 2025-09-07T08:12:53.4380124Z 2025-09-07T08:12:53.4380213Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4380395Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4380453Z return mod(**inputs) 2025-09-07T08:12:53.4380683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4380779Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4381012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4381158Z layer_outputs = layer_module( 2025-09-07T08:12:53.4381372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4381439Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4381666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4381733Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4381960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4382033Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4382264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4382381Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4382384Z 2025-09-07T08:12:53.4382473Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4382656Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4382711Z return mod(**inputs) 2025-09-07T08:12:53.4382941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4383002Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4383231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4383292Z layer_outputs = layer_module( 2025-09-07T08:12:53.4383503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4383574Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4383799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4383874Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4384099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4384172Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4384399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4384509Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4384512Z 2025-09-07T08:12:53.4384603Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4384819Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4384907Z return mod(**inputs) 2025-09-07T08:12:53.4385147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4385210Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4385442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4385506Z layer_outputs = layer_module( 2025-09-07T08:12:53.4385718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4385791Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4386015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4386091Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4386350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4386429Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4386687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4386799Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4386802Z 2025-09-07T08:12:53.4386894Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4387076Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4387132Z return mod(**inputs) 2025-09-07T08:12:53.4387365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4387429Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4387659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4387725Z layer_outputs = layer_module( 2025-09-07T08:12:53.4387937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4388004Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4388230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4388297Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4388522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4388593Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4388817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.4388961Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.4388967Z 2025-09-07T08:12:53.4389056Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4389237Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4389298Z return mod(**inputs) 2025-09-07T08:12:53.4389530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4389596Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4389826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4389889Z layer_outputs = layer_module( 2025-09-07T08:12:53.4390103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4390168Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4390431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4390532Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4390762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4390836Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4391066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.4391137Z value_states = self.v(current_states) 2025-09-07T08:12:53.4391140Z 2025-09-07T08:12:53.4391227Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4391416Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4391472Z return mod(**inputs) 2025-09-07T08:12:53.4391701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4391809Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4392040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4392103Z layer_outputs = layer_module( 2025-09-07T08:12:53.4392343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4392411Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4392638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4392706Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4392933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4393003Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4393233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4393328Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4393332Z 2025-09-07T08:12:53.4393419Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4393602Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4393656Z return mod(**inputs) 2025-09-07T08:12:53.4393887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4393947Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4394175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4394237Z layer_outputs = layer_module( 2025-09-07T08:12:53.4394446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4394517Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4394748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4394823Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4395051Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4395123Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4395355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4395445Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4395448Z 2025-09-07T08:12:53.4395536Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4395716Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4395804Z return mod(**inputs) 2025-09-07T08:12:53.4396069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4396131Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4396364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4396424Z layer_outputs = layer_module( 2025-09-07T08:12:53.4396635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4396702Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4396926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4396997Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4397222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4397328Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4397555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4397674Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4397678Z 2025-09-07T08:12:53.4397767Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4397948Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4398005Z return mod(**inputs) 2025-09-07T08:12:53.4398233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4398294Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4398524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4398588Z layer_outputs = layer_module( 2025-09-07T08:12:53.4398806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4398873Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4399108Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4399177Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4399401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4399475Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4399700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.4399796Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.4399799Z 2025-09-07T08:12:53.4399889Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4400075Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4400133Z return mod(**inputs) 2025-09-07T08:12:53.4400365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4400431Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4400657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4400716Z layer_outputs = layer_module( 2025-09-07T08:12:53.4400930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4400997Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4401228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4401336Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4401596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4401670Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4401896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.4401969Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.4401972Z 2025-09-07T08:12:53.4402042Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4402131Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4402311Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4402365Z return mod(**inputs) 2025-09-07T08:12:53.4402595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4402657Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4402921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4402982Z layer_outputs = layer_module( 2025-09-07T08:12:53.4403352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4403421Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4403656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4403742Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4403966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:12:53.4404052Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4404277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4404349Z return self.weight * hidden_states 2025-09-07T08:12:53.4404353Z 2025-09-07T08:12:53.4404445Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4404625Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4404681Z return mod(**inputs) 2025-09-07T08:12:53.4404912Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4404974Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4405202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4405262Z layer_outputs = layer_module( 2025-09-07T08:12:53.4405474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4405540Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4405769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4405846Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4406073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4406184Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4406415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:12:53.4406485Z hidden_states = self.wi(hidden_states) 2025-09-07T08:12:53.4406488Z 2025-09-07T08:12:53.4406575Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4406757Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4406860Z return mod(**inputs) 2025-09-07T08:12:53.4407094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4407202Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4407435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4407495Z layer_outputs = layer_module( 2025-09-07T08:12:53.4407707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4407772Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4407999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4408073Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4408300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4408405Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4408674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:12:53.4408748Z hidden_states = self.act(hidden_states) 2025-09-07T08:12:53.4408783Z 2025-09-07T08:12:53.4408871Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4409054Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4409108Z return mod(**inputs) 2025-09-07T08:12:53.4409335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4409397Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4409624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4409687Z layer_outputs = layer_module( 2025-09-07T08:12:53.4409899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4409965Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4410208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4410282Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4410509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4410606Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4410832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:12:53.4410899Z hidden_states = self.wo(hidden_states) 2025-09-07T08:12:53.4410903Z 2025-09-07T08:12:53.4410989Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4411177Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4411232Z return mod(**inputs) 2025-09-07T08:12:53.4411461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4411523Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4411753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4411815Z layer_outputs = layer_module( 2025-09-07T08:12:53.4412024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4412091Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4412320Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4412427Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4412703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 343, in forward 2025-09-07T08:12:53.4412822Z hidden_states = hidden_states + self.dropout(forwarded_states) 2025-09-07T08:12:53.4412825Z 2025-09-07T08:12:53.4412897Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4412985Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4413168Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4413222Z return mod(**inputs) 2025-09-07T08:12:53.4413452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4413515Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4413742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4413805Z layer_outputs = layer_module( 2025-09-07T08:12:53.4414047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4414112Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4414372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4414445Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4414674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:12:53.4414768Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4414994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4415059Z return self.weight * hidden_states 2025-09-07T08:12:53.4415063Z 2025-09-07T08:12:53.4415152Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4415337Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4415391Z return mod(**inputs) 2025-09-07T08:12:53.4415622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4415683Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4415909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4415970Z layer_outputs = layer_module( 2025-09-07T08:12:53.4416179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4416245Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4416470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4416541Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4416772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4416843Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4417070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.4417137Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.4417140Z 2025-09-07T08:12:53.4417228Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4417409Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4417463Z return mod(**inputs) 2025-09-07T08:12:53.4417751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4417812Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4418086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4418176Z layer_outputs = layer_module( 2025-09-07T08:12:53.4418386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4418455Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4418681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4418752Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4418977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4419046Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4419274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.4419342Z key_states = self.k(current_states) 2025-09-07T08:12:53.4419347Z 2025-09-07T08:12:53.4419471Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4419653Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4419709Z return mod(**inputs) 2025-09-07T08:12:53.4419968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4420031Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4420259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4420319Z layer_outputs = layer_module( 2025-09-07T08:12:53.4420530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4420594Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4420822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4420895Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4421119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4421191Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4421414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4421528Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4421534Z 2025-09-07T08:12:53.4421621Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4421801Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4421858Z return mod(**inputs) 2025-09-07T08:12:53.4422088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4422155Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4422381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4422441Z layer_outputs = layer_module( 2025-09-07T08:12:53.4422653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4422717Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4422944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4423010Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4423234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4423304Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4423568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4423712Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4423715Z 2025-09-07T08:12:53.4423804Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4423996Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4424050Z return mod(**inputs) 2025-09-07T08:12:53.4424277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4424342Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4424569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4424631Z layer_outputs = layer_module( 2025-09-07T08:12:53.4424840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4424938Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4425169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4425267Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4425498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4425567Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4425792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4425900Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4425904Z 2025-09-07T08:12:53.4425990Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4426175Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4426234Z return mod(**inputs) 2025-09-07T08:12:53.4426466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4426530Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4426760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4426824Z layer_outputs = layer_module( 2025-09-07T08:12:53.4427033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4427102Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4427328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4427397Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4427624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4427700Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4427931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.4428077Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.4428081Z 2025-09-07T08:12:53.4428175Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4428356Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4428411Z return mod(**inputs) 2025-09-07T08:12:53.4428651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4428714Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4428953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4429074Z layer_outputs = layer_module( 2025-09-07T08:12:53.4429288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4429357Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4429587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4429665Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4429890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4429958Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4430186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.4430256Z value_states = self.v(current_states) 2025-09-07T08:12:53.4430261Z 2025-09-07T08:12:53.4430356Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4430566Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4430625Z return mod(**inputs) 2025-09-07T08:12:53.4430884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4430947Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4431178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4431237Z layer_outputs = layer_module( 2025-09-07T08:12:53.4431449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4431517Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4431742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4431816Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4432041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4432115Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4432338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4432438Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4432446Z 2025-09-07T08:12:53.4432536Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4432718Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4432777Z return mod(**inputs) 2025-09-07T08:12:53.4433004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4433072Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4433301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4433362Z layer_outputs = layer_module( 2025-09-07T08:12:53.4433580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4433646Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4433875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4433946Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4434172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4434247Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4434472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4434645Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4434649Z 2025-09-07T08:12:53.4434737Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4434926Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4434981Z return mod(**inputs) 2025-09-07T08:12:53.4435209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4435274Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4435504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4435567Z layer_outputs = layer_module( 2025-09-07T08:12:53.4435779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4435847Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4436109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4436180Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4436444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4436513Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4436743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4436834Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4436838Z 2025-09-07T08:12:53.4436926Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4437111Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4437166Z return mod(**inputs) 2025-09-07T08:12:53.4437403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4437466Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4437695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4437756Z layer_outputs = layer_module( 2025-09-07T08:12:53.4437967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4438036Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4438267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4438335Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4438564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4438635Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4438864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.4438957Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.4438960Z 2025-09-07T08:12:53.4439052Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4439236Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4439290Z return mod(**inputs) 2025-09-07T08:12:53.4439522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4439584Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4439812Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4439873Z layer_outputs = layer_module( 2025-09-07T08:12:53.4440124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4440223Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4440451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:12:53.4440522Z self_attention_outputs = self.layer[0]( 2025-09-07T08:12:53.4440748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:12:53.4440817Z attention_output = self.SelfAttention( 2025-09-07T08:12:53.4441047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.4441113Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.4441117Z 2025-09-07T08:12:53.4441191Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4441280Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4441495Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4441557Z return mod(**inputs) 2025-09-07T08:12:53.4441785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4441983Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4442217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4442287Z layer_outputs = layer_module( 2025-09-07T08:12:53.4442500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4442566Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4442797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4442867Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4443107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T08:12:53.4443209Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4443437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4443511Z return self.weight * hidden_states 2025-09-07T08:12:53.4443514Z 2025-09-07T08:12:53.4443606Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4443798Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4443854Z return mod(**inputs) 2025-09-07T08:12:53.4444085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4444155Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4444388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4444458Z layer_outputs = layer_module( 2025-09-07T08:12:53.4444671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4444747Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4444978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4445048Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4445280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4445353Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4445584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:12:53.4445686Z query_states = self.q(hidden_states) 2025-09-07T08:12:53.4445717Z 2025-09-07T08:12:53.4445811Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4446000Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4446057Z return mod(**inputs) 2025-09-07T08:12:53.4446295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4446362Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4446599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4446663Z layer_outputs = layer_module( 2025-09-07T08:12:53.4446880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4446953Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4447213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4447292Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4447522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4447625Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4447859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:12:53.4447928Z key_states = self.k(current_states) 2025-09-07T08:12:53.4447931Z 2025-09-07T08:12:53.4448028Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4448212Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4448269Z return mod(**inputs) 2025-09-07T08:12:53.4448505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4448572Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4448805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4448864Z layer_outputs = layer_module( 2025-09-07T08:12:53.4449079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4449144Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4449368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4449437Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4449660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4449732Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4449958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4450072Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4450075Z 2025-09-07T08:12:53.4450164Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4450346Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4450403Z return mod(**inputs) 2025-09-07T08:12:53.4450630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4450692Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4450921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4450980Z layer_outputs = layer_module( 2025-09-07T08:12:53.4451191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4451319Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4451548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4451617Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4451843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4451916Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4452140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4452250Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4452253Z 2025-09-07T08:12:53.4452339Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4452519Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4452579Z return mod(**inputs) 2025-09-07T08:12:53.4452840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4452904Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4453176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4453239Z layer_outputs = layer_module( 2025-09-07T08:12:53.4453449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4453513Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4453740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4453807Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4454032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4454107Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4454329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:12:53.4454450Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:12:53.4454453Z 2025-09-07T08:12:53.4454540Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4454721Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4454776Z return mod(**inputs) 2025-09-07T08:12:53.4455003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4455068Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4455299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4455369Z layer_outputs = layer_module( 2025-09-07T08:12:53.4455585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4455657Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4455888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4455959Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4456193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4456264Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4456498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:12:53.4456640Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:12:53.4456678Z 2025-09-07T08:12:53.4456795Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4456985Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4457042Z return mod(**inputs) 2025-09-07T08:12:53.4457281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4457346Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4457638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4457701Z layer_outputs = layer_module( 2025-09-07T08:12:53.4457914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4457985Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4458208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4458318Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4458549Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4458622Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4458886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:12:53.4458956Z value_states = self.v(current_states) 2025-09-07T08:12:53.4458959Z 2025-09-07T08:12:53.4459051Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4459234Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4459291Z return mod(**inputs) 2025-09-07T08:12:53.4459523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4459587Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4459826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4459888Z layer_outputs = layer_module( 2025-09-07T08:12:53.4460108Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4460176Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4460403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4460478Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4460705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4460782Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4461008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4461106Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4461110Z 2025-09-07T08:12:53.4461205Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4461386Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4461452Z return mod(**inputs) 2025-09-07T08:12:53.4461681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4461744Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4461978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4462040Z layer_outputs = layer_module( 2025-09-07T08:12:53.4462258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4462358Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4462624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4462703Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4462930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4463005Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4463232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4463331Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4463334Z 2025-09-07T08:12:53.4463424Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4463606Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4463668Z return mod(**inputs) 2025-09-07T08:12:53.4463927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4463999Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4464260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4464322Z layer_outputs = layer_module( 2025-09-07T08:12:53.4464558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4464627Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4464856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4464926Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4465160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4465235Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4465464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:12:53.4465560Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:12:53.4465563Z 2025-09-07T08:12:53.4465653Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4465841Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4465897Z return mod(**inputs) 2025-09-07T08:12:53.4466128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4466195Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4466425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4466497Z layer_outputs = layer_module( 2025-09-07T08:12:53.4466709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4466779Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4467006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4467074Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4467304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4467373Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4467602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:12:53.4467697Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:12:53.4467700Z 2025-09-07T08:12:53.4467789Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4468010Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4468099Z return mod(**inputs) 2025-09-07T08:12:53.4468332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4468395Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4468626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4468692Z layer_outputs = layer_module( 2025-09-07T08:12:53.4468907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4468978Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4469202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:12:53.4469280Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:12:53.4469587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:12:53.4469661Z attention_output = self.EncDecAttention( 2025-09-07T08:12:53.4469925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:12:53.4469995Z attn_output = self.o(attn_output) 2025-09-07T08:12:53.4469998Z 2025-09-07T08:12:53.4470074Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4470162Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4470343Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4470403Z return mod(**inputs) 2025-09-07T08:12:53.4470630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4470699Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4470927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4470990Z layer_outputs = layer_module( 2025-09-07T08:12:53.4471207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4471273Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4471500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4471584Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4471816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:12:53.4471901Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:12:53.4472130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:12:53.4472206Z return self.weight * hidden_states 2025-09-07T08:12:53.4472211Z 2025-09-07T08:12:53.4472302Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4472485Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4472542Z return mod(**inputs) 2025-09-07T08:12:53.4472788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4472859Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4473093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4473159Z layer_outputs = layer_module( 2025-09-07T08:12:53.4473371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4473438Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4473706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4473815Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4474047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4474154Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4474385Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:12:53.4474456Z hidden_states = self.wi(hidden_states) 2025-09-07T08:12:53.4474460Z 2025-09-07T08:12:53.4474550Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4474735Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4474790Z return mod(**inputs) 2025-09-07T08:12:53.4475069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4475135Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4475368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4475460Z layer_outputs = layer_module( 2025-09-07T08:12:53.4475676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4475751Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4475977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4476058Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4476286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4476387Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4476621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:12:53.4476695Z hidden_states = self.act(hidden_states) 2025-09-07T08:12:53.4476698Z 2025-09-07T08:12:53.4476790Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4476970Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4477027Z return mod(**inputs) 2025-09-07T08:12:53.4477262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:12:53.4477324Z decoder_outputs = self.decoder( 2025-09-07T08:12:53.4477555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:12:53.4477617Z layer_outputs = layer_module( 2025-09-07T08:12:53.4477831Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:12:53.4477905Z return super().__call__(*args, **kwargs) 2025-09-07T08:12:53.4478133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:12:53.4478214Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:12:53.4478439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:12:53.4478545Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:12:53.4478771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:12:53.4478840Z hidden_states = self.wo(hidden_states) 2025-09-07T08:12:53.4478844Z 2025-09-07T08:12:53.4478919Z cudagraph partition due to non gpu ops 2025-09-07T08:12:53.4479038Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4479256Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4479313Z return mod(**inputs) 2025-09-07T08:12:53.4479545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1789, in forward 2025-09-07T08:12:53.4479658Z sequence_output = sequence_output * (self.model_dim**-0.5) 2025-09-07T08:12:53.4479661Z 2025-09-07T08:12:53.4479749Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:12:53.4479939Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:12:53.4479995Z return mod(**inputs) 2025-09-07T08:12:53.4480224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1791, in forward 2025-09-07T08:12:53.4480305Z lm_logits = self.lm_head(sequence_output) 2025-09-07T08:12:53.4480309Z 2025-09-07T08:13:14.2770757Z pass 2025-09-07T08:13:14.2771533Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:13:17.1665731Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:13:17.1669222Z import pynvml # type: ignore[import] 2025-09-07T08:13:19.1728476Z 2025-09-07T08:13:28.6664800Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:13:28.6665064Z loading model: 0it [00:09, ?it/s] 2025-09-07T08:13:28.7036056Z cpu eval hf_T5_base 2025-09-07T08:13:43.0941899Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:13:49.6123556Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:13:56.1018394Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:14:18.8250411Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.8250744Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8251164Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8251506Z return mod(**inputs) 2025-09-07T08:14:18.8251899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.8252305Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.8252669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8253042Z layer_outputs = layer_module( 2025-09-07T08:14:18.8253394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8253768Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8254156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8254532Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8254911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8255293Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8255663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 546, in forward 2025-09-07T08:14:18.8256075Z position_bias = position_bias + causal_mask 2025-09-07T08:14:18.8256217Z 2025-09-07T08:14:18.8256334Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8256730Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8257525Z return mod(**inputs) 2025-09-07T08:14:18.8257878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.8258384Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.8258763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8259140Z layer_outputs = layer_module( 2025-09-07T08:14:18.8259487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8259856Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8260248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8260619Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8260990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.8261400Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.8261911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.8262292Z return self.weight * hidden_states 2025-09-07T08:14:18.8262423Z 2025-09-07T08:14:18.8262602Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8262969Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8263311Z return mod(**inputs) 2025-09-07T08:14:18.8263697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.8264100Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.8264472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8264849Z layer_outputs = layer_module( 2025-09-07T08:14:18.8265206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8265574Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8265950Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8266334Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8266708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8267093Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8267471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.8267839Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.8267979Z 2025-09-07T08:14:18.8268084Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8276786Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8277186Z return mod(**inputs) 2025-09-07T08:14:18.8277587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.8278002Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.8278403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8278836Z layer_outputs = layer_module( 2025-09-07T08:14:18.8279362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8279740Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8280131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8280521Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8280974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8281400Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8281780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.8282159Z key_states = self.k(current_states) 2025-09-07T08:14:18.8282290Z 2025-09-07T08:14:18.8282403Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8282769Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8283102Z return mod(**inputs) 2025-09-07T08:14:18.8283452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.8283837Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.8284204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8284614Z layer_outputs = layer_module( 2025-09-07T08:14:18.8284964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8285332Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8285741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8286129Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8286488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8286870Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8287231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8287657Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8287845Z 2025-09-07T08:14:18.8287957Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8288332Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8288656Z return mod(**inputs) 2025-09-07T08:14:18.8289021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.8289396Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.8289748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8290115Z layer_outputs = layer_module( 2025-09-07T08:14:18.8290459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8290821Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8291185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8291563Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8291928Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8292310Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8292673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8293077Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8293266Z 2025-09-07T08:14:18.8293362Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8293718Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8294032Z return mod(**inputs) 2025-09-07T08:14:18.8294364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.8294812Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.8295183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8295552Z layer_outputs = layer_module( 2025-09-07T08:14:18.8295888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8296241Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8296600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8296970Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8297394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8297776Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8298132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8298593Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8298780Z 2025-09-07T08:14:18.8298877Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8299273Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8299595Z return mod(**inputs) 2025-09-07T08:14:18.8299926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.8300298Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.8300659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8301026Z layer_outputs = layer_module( 2025-09-07T08:14:18.8301367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8301723Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8302094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8302464Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8302835Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8303377Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8303745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8304197Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8304412Z 2025-09-07T08:14:18.8304510Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8304865Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8305182Z return mod(**inputs) 2025-09-07T08:14:18.8305529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.8305901Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.8306261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8306626Z layer_outputs = layer_module( 2025-09-07T08:14:18.8306954Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8307311Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8307676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8308044Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8308397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8308912Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8309294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.8309669Z value_states = self.v(current_states) 2025-09-07T08:14:18.8309798Z 2025-09-07T08:14:18.8309905Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8310257Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8310578Z return mod(**inputs) 2025-09-07T08:14:18.8310934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.8311310Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.8311672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8312033Z layer_outputs = layer_module( 2025-09-07T08:14:18.8312427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8312798Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8313220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8313593Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8313957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8314338Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8314703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8315103Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8315271Z 2025-09-07T08:14:18.8315371Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8315719Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8316038Z return mod(**inputs) 2025-09-07T08:14:18.8316381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.8316755Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.8317101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8317467Z layer_outputs = layer_module( 2025-09-07T08:14:18.8317807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8318163Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8318521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8318893Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8319260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8319636Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8320000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8320390Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8320552Z 2025-09-07T08:14:18.8320644Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8320995Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8321309Z return mod(**inputs) 2025-09-07T08:14:18.8321647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.8322005Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.8322405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8322825Z layer_outputs = layer_module( 2025-09-07T08:14:18.8323155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8323516Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8323875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8324251Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8324611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8324983Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8325334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8325726Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8325952Z 2025-09-07T08:14:18.8326046Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8326403Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8326750Z return mod(**inputs) 2025-09-07T08:14:18.8327092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.8327476Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.8327830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8328195Z layer_outputs = layer_module( 2025-09-07T08:14:18.8328518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8328876Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8329236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8329613Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8329968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8330331Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8330688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.8331083Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.8331235Z 2025-09-07T08:14:18.8331336Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8331680Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8331990Z return mod(**inputs) 2025-09-07T08:14:18.8332329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.8332698Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.8333048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8333403Z layer_outputs = layer_module( 2025-09-07T08:14:18.8333748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8334100Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8334462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8334823Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8335181Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8335552Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8335961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.8336367Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.8336492Z 2025-09-07T08:14:18.8336586Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8336934Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8337306Z return mod(**inputs) 2025-09-07T08:14:18.8337644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.8338011Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.8338360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8338725Z layer_outputs = layer_module( 2025-09-07T08:14:18.8339059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8339458Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8339821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.8340226Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.8340608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.8340985Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.8341346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.8341711Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.8341845Z 2025-09-07T08:14:18.8341935Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8342273Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8342591Z return mod(**inputs) 2025-09-07T08:14:18.8342924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8343305Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8343659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8344027Z layer_outputs = layer_module( 2025-09-07T08:14:18.8344356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8344695Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8345062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8345438Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8345805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8346186Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8346539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.8346910Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.8347039Z 2025-09-07T08:14:18.8347138Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8347479Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8347792Z return mod(**inputs) 2025-09-07T08:14:18.8348124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8348490Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8348848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8350502Z layer_outputs = layer_module( 2025-09-07T08:14:18.8350876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8351239Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8351635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8352010Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8352370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8352741Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8353125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.8353494Z key_states = self.k(current_states) 2025-09-07T08:14:18.8353634Z 2025-09-07T08:14:18.8353744Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8354127Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8354446Z return mod(**inputs) 2025-09-07T08:14:18.8354791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8355191Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8355561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8355927Z layer_outputs = layer_module( 2025-09-07T08:14:18.8356261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8356633Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8356997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8357377Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8357734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8358106Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8358475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8358895Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8359074Z 2025-09-07T08:14:18.8359177Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8359519Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8359831Z return mod(**inputs) 2025-09-07T08:14:18.8360163Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8360533Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8360888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8361249Z layer_outputs = layer_module( 2025-09-07T08:14:18.8361588Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8361947Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8362308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8362674Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8363033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8363409Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8363768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8364225Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8364443Z 2025-09-07T08:14:18.8364536Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8364882Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8365202Z return mod(**inputs) 2025-09-07T08:14:18.8365530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8365899Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8366252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8366609Z layer_outputs = layer_module( 2025-09-07T08:14:18.8366945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8367294Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8367683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8368059Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8368452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8368831Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8369186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8369593Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8369776Z 2025-09-07T08:14:18.8369869Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8370220Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8370528Z return mod(**inputs) 2025-09-07T08:14:18.8370853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8371227Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8371580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8371946Z layer_outputs = layer_module( 2025-09-07T08:14:18.8372278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8372624Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8372985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8373352Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8373713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8374076Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8374457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8374909Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8375119Z 2025-09-07T08:14:18.8375221Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8375558Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8375885Z return mod(**inputs) 2025-09-07T08:14:18.8376214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8376574Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8376925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8377326Z layer_outputs = layer_module( 2025-09-07T08:14:18.8377708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8378114Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8378481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8378850Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8379207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8379574Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8379934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8380376Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8380576Z 2025-09-07T08:14:18.8380676Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8381018Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8381372Z return mod(**inputs) 2025-09-07T08:14:18.8381704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8382100Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8382461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8382823Z layer_outputs = layer_module( 2025-09-07T08:14:18.8383160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8383508Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8383870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8384245Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8384598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8384978Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8385379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.8385753Z value_states = self.v(current_states) 2025-09-07T08:14:18.8385889Z 2025-09-07T08:14:18.8385984Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8386326Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8386644Z return mod(**inputs) 2025-09-07T08:14:18.8386974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8387355Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8387696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8388066Z layer_outputs = layer_module( 2025-09-07T08:14:18.8388397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8388751Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8389105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8389484Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8389842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8390235Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8390596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8390998Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8391208Z 2025-09-07T08:14:18.8391332Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8391676Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8391995Z return mod(**inputs) 2025-09-07T08:14:18.8392325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8392693Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8393042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8393408Z layer_outputs = layer_module( 2025-09-07T08:14:18.8393744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8394091Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8394454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8394862Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8395225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8395635Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8395989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8396387Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8396547Z 2025-09-07T08:14:18.8396643Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8396981Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8397297Z return mod(**inputs) 2025-09-07T08:14:18.8397619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8397984Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8398341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8398705Z layer_outputs = layer_module( 2025-09-07T08:14:18.8399034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8399382Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8399744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8400107Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8400471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8400830Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8401211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8401603Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8401756Z 2025-09-07T08:14:18.8401864Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8402209Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8402531Z return mod(**inputs) 2025-09-07T08:14:18.8402875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8403478Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8403843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8404219Z layer_outputs = layer_module( 2025-09-07T08:14:18.8404559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8404994Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8405415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8405798Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8406167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8406542Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8406916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.8407318Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.8407480Z 2025-09-07T08:14:18.8407587Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8407936Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8408264Z return mod(**inputs) 2025-09-07T08:14:18.8408653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8409030Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8409443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8409822Z layer_outputs = layer_module( 2025-09-07T08:14:18.8410165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8410534Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8410906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8411281Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8411655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8412050Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8412440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.8412808Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.8412935Z 2025-09-07T08:14:18.8413019Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.8413252Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8413604Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8413923Z return mod(**inputs) 2025-09-07T08:14:18.8414263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8414639Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8414994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8415364Z layer_outputs = layer_module( 2025-09-07T08:14:18.8415703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8416057Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8416417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8416804Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8417289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.8417679Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.8418069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.8418438Z return self.weight * hidden_states 2025-09-07T08:14:18.8418614Z 2025-09-07T08:14:18.8418722Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8419110Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8419434Z return mod(**inputs) 2025-09-07T08:14:18.8419783Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8420153Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8420537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8420902Z layer_outputs = layer_module( 2025-09-07T08:14:18.8421239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8421601Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8421959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8422357Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8422763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8423181Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8423625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.8423998Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.8424125Z 2025-09-07T08:14:18.8424226Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8424570Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8424890Z return mod(**inputs) 2025-09-07T08:14:18.8425234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8425598Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8425954Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8426316Z layer_outputs = layer_module( 2025-09-07T08:14:18.8426654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8427013Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8427375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8427756Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8428128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8428546Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8428942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.8429323Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.8429450Z 2025-09-07T08:14:18.8429545Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8429895Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8430207Z return mod(**inputs) 2025-09-07T08:14:18.8430546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8430908Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8431253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8431617Z layer_outputs = layer_module( 2025-09-07T08:14:18.8431952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8432369Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8432763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8433139Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8433516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8433918Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8434309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.8434676Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.8434809Z 2025-09-07T08:14:18.8434884Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.8435107Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8435449Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8435762Z return mod(**inputs) 2025-09-07T08:14:18.8436127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8436498Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8436889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8437257Z layer_outputs = layer_module( 2025-09-07T08:14:18.8437587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8437942Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8438307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8438683Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8439044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.8439437Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.8439825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.8440192Z return self.weight * hidden_states 2025-09-07T08:14:18.8440321Z 2025-09-07T08:14:18.8440424Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8440768Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8441075Z return mod(**inputs) 2025-09-07T08:14:18.8441403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8441771Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8442125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8442486Z layer_outputs = layer_module( 2025-09-07T08:14:18.8442821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8443172Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8443532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8443901Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8444254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8444622Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8444987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.8445350Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.8445477Z 2025-09-07T08:14:18.8445571Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8445995Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8446312Z return mod(**inputs) 2025-09-07T08:14:18.8446641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8447005Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8447352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8447716Z layer_outputs = layer_module( 2025-09-07T08:14:18.8448045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8448413Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8448787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8449160Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8449563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8449953Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8450355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.8450720Z key_states = self.k(current_states) 2025-09-07T08:14:18.8450859Z 2025-09-07T08:14:18.8450960Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8451324Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8451646Z return mod(**inputs) 2025-09-07T08:14:18.8451987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8452353Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8452719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8453087Z layer_outputs = layer_module( 2025-09-07T08:14:18.8453430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8453779Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8454141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8454507Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8454870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8455241Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8455595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8456011Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8456201Z 2025-09-07T08:14:18.8456300Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8456649Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8456964Z return mod(**inputs) 2025-09-07T08:14:18.8457352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8457722Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8458073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8458438Z layer_outputs = layer_module( 2025-09-07T08:14:18.8458767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8459122Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8459533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8459944Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8460307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8460671Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8461030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8461438Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8461614Z 2025-09-07T08:14:18.8461717Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8462066Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8462374Z return mod(**inputs) 2025-09-07T08:14:18.8462711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8463110Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8463465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8463855Z layer_outputs = layer_module( 2025-09-07T08:14:18.8464202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8464577Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8464938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8465306Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8465659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8466031Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8466407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8466817Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8466990Z 2025-09-07T08:14:18.8467091Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8467425Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8467737Z return mod(**inputs) 2025-09-07T08:14:18.8468070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8468436Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8468782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8469141Z layer_outputs = layer_module( 2025-09-07T08:14:18.8469475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8469830Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8470187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8470550Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8470919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8471286Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8471758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8472200Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8472412Z 2025-09-07T08:14:18.8472506Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8472905Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8473264Z return mod(**inputs) 2025-09-07T08:14:18.8473604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8473966Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8474321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8474683Z layer_outputs = layer_module( 2025-09-07T08:14:18.8475017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8475369Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8475722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8476091Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8476493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8476875Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8477236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8477722Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8477939Z 2025-09-07T08:14:18.8478038Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8478385Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8478699Z return mod(**inputs) 2025-09-07T08:14:18.8479026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8479392Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8479744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8480120Z layer_outputs = layer_module( 2025-09-07T08:14:18.8480462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8480822Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8481194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8481567Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8481955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8482321Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8482685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.8483050Z value_states = self.v(current_states) 2025-09-07T08:14:18.8483181Z 2025-09-07T08:14:18.8483292Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8483646Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8483958Z return mod(**inputs) 2025-09-07T08:14:18.8484300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8484667Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8485025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8485384Z layer_outputs = layer_module( 2025-09-07T08:14:18.8485727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8486078Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8486442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8486893Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8487246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8487621Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8487989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8488404Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8488563Z 2025-09-07T08:14:18.8488667Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8489007Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8489322Z return mod(**inputs) 2025-09-07T08:14:18.8489656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8490023Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8490413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8490783Z layer_outputs = layer_module( 2025-09-07T08:14:18.8491150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8491507Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8491868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8492231Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8492591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8492968Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8493345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8493756Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8493915Z 2025-09-07T08:14:18.8494016Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8494373Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8494704Z return mod(**inputs) 2025-09-07T08:14:18.8495053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8495419Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8495782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8496148Z layer_outputs = layer_module( 2025-09-07T08:14:18.8496497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8496860Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8497264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8497641Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8498008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8498382Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8498742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8499141Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8499302Z 2025-09-07T08:14:18.8499398Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8499750Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8500122Z return mod(**inputs) 2025-09-07T08:14:18.8500493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8500864Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8501219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8501585Z layer_outputs = layer_module( 2025-09-07T08:14:18.8501923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8502269Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8502633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8503005Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8503512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8503880Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8504297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.8504694Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.8504897Z 2025-09-07T08:14:18.8504997Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8505344Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8505665Z return mod(**inputs) 2025-09-07T08:14:18.8506007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8506382Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8506746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8507115Z layer_outputs = layer_module( 2025-09-07T08:14:18.8507452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8507810Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8508178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8508556Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8508915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8509285Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8526412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.8526819Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.8526964Z 2025-09-07T08:14:18.8527073Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8527445Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8527776Z return mod(**inputs) 2025-09-07T08:14:18.8528142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8528567Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8528944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8529303Z layer_outputs = layer_module( 2025-09-07T08:14:18.8529648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8530009Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8530380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8530771Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8531259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.8531714Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.8532098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.8532460Z return self.weight * hidden_states 2025-09-07T08:14:18.8532586Z 2025-09-07T08:14:18.8532691Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8533050Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8533380Z return mod(**inputs) 2025-09-07T08:14:18.8533739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8534108Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8534464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8534869Z layer_outputs = layer_module( 2025-09-07T08:14:18.8535208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8535996Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8536364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8536740Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8537111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8537606Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8537998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.8538362Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.8538495Z 2025-09-07T08:14:18.8538593Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8538946Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8539262Z return mod(**inputs) 2025-09-07T08:14:18.8539598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8539959Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8540307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8540661Z layer_outputs = layer_module( 2025-09-07T08:14:18.8540991Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8541336Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8541692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8542065Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8542429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8542825Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8543214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.8543578Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.8543707Z 2025-09-07T08:14:18.8543801Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8544140Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8544476Z return mod(**inputs) 2025-09-07T08:14:18.8544819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8545267Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8545623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8545978Z layer_outputs = layer_module( 2025-09-07T08:14:18.8546307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8546652Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8547017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8547405Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8547771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8548168Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8548606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.8548973Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.8549103Z 2025-09-07T08:14:18.8549178Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.8549432Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8549784Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8550095Z return mod(**inputs) 2025-09-07T08:14:18.8550431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8550793Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8551141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8551495Z layer_outputs = layer_module( 2025-09-07T08:14:18.8551829Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8552182Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8552533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8552898Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8553247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.8553636Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.8554018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.8554373Z return self.weight * hidden_states 2025-09-07T08:14:18.8554496Z 2025-09-07T08:14:18.8554597Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8554937Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8555250Z return mod(**inputs) 2025-09-07T08:14:18.8555574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8555932Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8556279Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8556630Z layer_outputs = layer_module( 2025-09-07T08:14:18.8556955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8557299Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8557650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8558009Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8558403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8558803Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8559161Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.8559520Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.8559646Z 2025-09-07T08:14:18.8559737Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8560074Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8560379Z return mod(**inputs) 2025-09-07T08:14:18.8560700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8561051Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8561397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8561786Z layer_outputs = layer_module( 2025-09-07T08:14:18.8562110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8562457Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8562838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8563216Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8563583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8563953Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8564307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.8564663Z key_states = self.k(current_states) 2025-09-07T08:14:18.8564798Z 2025-09-07T08:14:18.8564894Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8565246Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8565565Z return mod(**inputs) 2025-09-07T08:14:18.8565887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8566245Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8566589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8566941Z layer_outputs = layer_module( 2025-09-07T08:14:18.8567265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8567607Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8567956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8568316Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8568665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8569029Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8569383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8569798Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8570001Z 2025-09-07T08:14:18.8570096Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8570435Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8570745Z return mod(**inputs) 2025-09-07T08:14:18.8571090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8571501Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8571878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8572232Z layer_outputs = layer_module( 2025-09-07T08:14:18.8572558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8572908Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8573261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8573628Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8573976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8574335Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8574685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8575125Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8575297Z 2025-09-07T08:14:18.8575391Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8575755Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8576066Z return mod(**inputs) 2025-09-07T08:14:18.8576391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8576747Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8577088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8577502Z layer_outputs = layer_module( 2025-09-07T08:14:18.8577827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8578172Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8578522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8578881Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8579231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8579595Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8579949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8580353Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8580526Z 2025-09-07T08:14:18.8580619Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8580973Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8581291Z return mod(**inputs) 2025-09-07T08:14:18.8581621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8581982Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8582325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8582681Z layer_outputs = layer_module( 2025-09-07T08:14:18.8583002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8583348Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8583703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8584061Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8584416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8584835Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8585227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8585674Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8585881Z 2025-09-07T08:14:18.8585974Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8586312Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8586626Z return mod(**inputs) 2025-09-07T08:14:18.8586949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8587313Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8587658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8588024Z layer_outputs = layer_module( 2025-09-07T08:14:18.8588384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8588741Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8589120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8589489Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8589837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8590199Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8590564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8590988Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8591200Z 2025-09-07T08:14:18.8591301Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8591637Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8591949Z return mod(**inputs) 2025-09-07T08:14:18.8592269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8592623Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8592972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8593322Z layer_outputs = layer_module( 2025-09-07T08:14:18.8593646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8593986Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8594341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8594711Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8595073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8595431Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8595787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.8596145Z value_states = self.v(current_states) 2025-09-07T08:14:18.8596272Z 2025-09-07T08:14:18.8596368Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8596708Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8597008Z return mod(**inputs) 2025-09-07T08:14:18.8597334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8597694Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8598082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8598492Z layer_outputs = layer_module( 2025-09-07T08:14:18.8598819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8599168Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8599515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8599880Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8600226Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8600589Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8600943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8601343Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8601534Z 2025-09-07T08:14:18.8601629Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8601962Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8602300Z return mod(**inputs) 2025-09-07T08:14:18.8602622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8602994Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8603509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8603865Z layer_outputs = layer_module( 2025-09-07T08:14:18.8604199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8604543Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8604904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8605265Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8605616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8605983Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8606333Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8606735Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8606886Z 2025-09-07T08:14:18.8606976Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8607312Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8607623Z return mod(**inputs) 2025-09-07T08:14:18.8607950Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8608318Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8608661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8609028Z layer_outputs = layer_module( 2025-09-07T08:14:18.8609349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8609698Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8610044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8610409Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8610764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8611121Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8611556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8612000Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8612167Z 2025-09-07T08:14:18.8612265Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8612607Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8612927Z return mod(**inputs) 2025-09-07T08:14:18.8613271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8613645Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8614002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8614369Z layer_outputs = layer_module( 2025-09-07T08:14:18.8614705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8615101Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8615459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8615875Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8616235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8616597Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8616967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.8617425Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.8617584Z 2025-09-07T08:14:18.8617686Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8618042Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8618358Z return mod(**inputs) 2025-09-07T08:14:18.8618699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8619062Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8619423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8619778Z layer_outputs = layer_module( 2025-09-07T08:14:18.8620112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8620471Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8620826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8621191Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8621543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8621913Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8622276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.8622635Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.8622761Z 2025-09-07T08:14:18.8622849Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.8623064Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8623409Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8623726Z return mod(**inputs) 2025-09-07T08:14:18.8624060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8624426Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8624776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8625234Z layer_outputs = layer_module( 2025-09-07T08:14:18.8625563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8625926Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8626276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8626654Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8627026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.8627419Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.8627799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.8628160Z return self.weight * hidden_states 2025-09-07T08:14:18.8628289Z 2025-09-07T08:14:18.8628427Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8628768Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8629083Z return mod(**inputs) 2025-09-07T08:14:18.8629440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8629819Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8630172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8630535Z layer_outputs = layer_module( 2025-09-07T08:14:18.8630863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8631207Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8631568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8631944Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8632315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8632718Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8633113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.8633487Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.8633616Z 2025-09-07T08:14:18.8633715Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8634070Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8634384Z return mod(**inputs) 2025-09-07T08:14:18.8634715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8635091Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8635462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8635844Z layer_outputs = layer_module( 2025-09-07T08:14:18.8636180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8636541Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8636899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8637274Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8637651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8638054Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8638488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.8638904Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.8639049Z 2025-09-07T08:14:18.8639152Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8639497Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8639816Z return mod(**inputs) 2025-09-07T08:14:18.8640145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8640523Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8640878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8641237Z layer_outputs = layer_module( 2025-09-07T08:14:18.8641571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8641960Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8642340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8642708Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8643131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8643529Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8643933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.8644296Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.8644421Z 2025-09-07T08:14:18.8644493Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.8644711Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8645053Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8645367Z return mod(**inputs) 2025-09-07T08:14:18.8645691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8646051Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8646400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8646752Z layer_outputs = layer_module( 2025-09-07T08:14:18.8647076Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8647418Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8647775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8648139Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8648495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.8648890Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.8649267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.8649632Z return self.weight * hidden_states 2025-09-07T08:14:18.8649763Z 2025-09-07T08:14:18.8649856Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8650211Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8650518Z return mod(**inputs) 2025-09-07T08:14:18.8650852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8651215Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8651574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8652028Z layer_outputs = layer_module( 2025-09-07T08:14:18.8652355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8652708Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8653066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8653433Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8653791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8654174Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8654530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.8654894Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.8655037Z 2025-09-07T08:14:18.8655138Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8655511Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8655836Z return mod(**inputs) 2025-09-07T08:14:18.8656201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8656580Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8656929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8657320Z layer_outputs = layer_module( 2025-09-07T08:14:18.8657659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8658011Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8658383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8658755Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8659115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8659489Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8659853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.8660219Z key_states = self.k(current_states) 2025-09-07T08:14:18.8660345Z 2025-09-07T08:14:18.8660438Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8660773Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8661085Z return mod(**inputs) 2025-09-07T08:14:18.8661426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8661789Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8662154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8662513Z layer_outputs = layer_module( 2025-09-07T08:14:18.8662843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8663200Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8663556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8663931Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8664281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8664650Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8665013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8665511Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8665696Z 2025-09-07T08:14:18.8665787Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8666132Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8666451Z return mod(**inputs) 2025-09-07T08:14:18.8666797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8667160Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8667508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8667880Z layer_outputs = layer_module( 2025-09-07T08:14:18.8668205Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8668548Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8668955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8669322Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8669729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8670101Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8670456Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8670864Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8671061Z 2025-09-07T08:14:18.8671154Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8671511Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8671819Z return mod(**inputs) 2025-09-07T08:14:18.8672165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8672533Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8672877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8673254Z layer_outputs = layer_module( 2025-09-07T08:14:18.8673575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8673972Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8674329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8674690Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8675061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8675424Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8675803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8676227Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8676401Z 2025-09-07T08:14:18.8676499Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8676831Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8677147Z return mod(**inputs) 2025-09-07T08:14:18.8677476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8677845Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8678188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8678632Z layer_outputs = layer_module( 2025-09-07T08:14:18.8678972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8679375Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8679729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8680107Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8680456Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8680817Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8681172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8681632Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8681835Z 2025-09-07T08:14:18.8681925Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8682317Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8682669Z return mod(**inputs) 2025-09-07T08:14:18.8682997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8683399Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8683750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8684108Z layer_outputs = layer_module( 2025-09-07T08:14:18.8684445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8684790Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8685142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8685509Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8685879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8686243Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8686603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8687039Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8687246Z 2025-09-07T08:14:18.8687336Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8687674Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8687988Z return mod(**inputs) 2025-09-07T08:14:18.8688312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8688684Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8689043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8689401Z layer_outputs = layer_module( 2025-09-07T08:14:18.8689745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8690108Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8690469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8690847Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8691206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8691580Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8691926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.8692327Z value_states = self.v(current_states) 2025-09-07T08:14:18.8692504Z 2025-09-07T08:14:18.8692596Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8692936Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8693245Z return mod(**inputs) 2025-09-07T08:14:18.8693572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8693931Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8694282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8694645Z layer_outputs = layer_module( 2025-09-07T08:14:18.8694970Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8695319Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8695709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8696086Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8696491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8696863Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8697301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8697701Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8697859Z 2025-09-07T08:14:18.8697963Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8698304Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8698612Z return mod(**inputs) 2025-09-07T08:14:18.8698939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8699316Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8699666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8700021Z layer_outputs = layer_module( 2025-09-07T08:14:18.8700349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8700705Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8701061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8701430Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8701782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8702168Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8702530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8702918Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8703205Z 2025-09-07T08:14:18.8703303Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8703639Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8703955Z return mod(**inputs) 2025-09-07T08:14:18.8704284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8704648Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8704987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8705353Z layer_outputs = layer_module( 2025-09-07T08:14:18.8705684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8706149Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8706526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8706894Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8707266Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8707634Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8708013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8708401Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8708566Z 2025-09-07T08:14:18.8708665Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8709010Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8709340Z return mod(**inputs) 2025-09-07T08:14:18.8709735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8710107Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8710512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8710883Z layer_outputs = layer_module( 2025-09-07T08:14:18.8711241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8711599Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8711961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8712331Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8712691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8713060Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8713415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.8713812Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.8713971Z 2025-09-07T08:14:18.8714068Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8714413Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8714720Z return mod(**inputs) 2025-09-07T08:14:18.8715045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8715401Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8715745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8716115Z layer_outputs = layer_module( 2025-09-07T08:14:18.8716439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8716787Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8717146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8717512Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8717875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8718235Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8718588Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.8718947Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.8719116Z 2025-09-07T08:14:18.8719216Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8719597Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8719902Z return mod(**inputs) 2025-09-07T08:14:18.8720239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8720600Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8720946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8721301Z layer_outputs = layer_module( 2025-09-07T08:14:18.8721629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8721981Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8722340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8722712Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8723103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:14:18.8723528Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:14:18.8723751Z 2025-09-07T08:14:18.8723829Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.8724054Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8724386Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8724700Z return mod(**inputs) 2025-09-07T08:14:18.8725037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8725395Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8725744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8726105Z layer_outputs = layer_module( 2025-09-07T08:14:18.8726431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8726777Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8727140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8727517Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8727885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.8728271Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.8728648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.8729008Z return self.weight * hidden_states 2025-09-07T08:14:18.8729135Z 2025-09-07T08:14:18.8729228Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8729566Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8729875Z return mod(**inputs) 2025-09-07T08:14:18.8730203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8730562Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8730904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8731267Z layer_outputs = layer_module( 2025-09-07T08:14:18.8731599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8731955Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8732311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8732785Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8733162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8733567Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8733971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.8734332Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.8734465Z 2025-09-07T08:14:18.8734561Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8734910Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8735224Z return mod(**inputs) 2025-09-07T08:14:18.8735558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8735924Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8736323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8736691Z layer_outputs = layer_module( 2025-09-07T08:14:18.8737061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8737488Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8737856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8738233Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8738604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8739011Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8739404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.8739775Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.8739911Z 2025-09-07T08:14:18.8740005Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8740358Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8740667Z return mod(**inputs) 2025-09-07T08:14:18.8740995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8741359Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8741712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8742079Z layer_outputs = layer_module( 2025-09-07T08:14:18.8742406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8742787Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8743171Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8743554Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8743926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8744321Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8744718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.8745086Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.8745215Z 2025-09-07T08:14:18.8745297Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.8745529Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8745871Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8746264Z return mod(**inputs) 2025-09-07T08:14:18.8746613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8746982Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8747333Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8747697Z layer_outputs = layer_module( 2025-09-07T08:14:18.8748038Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8748402Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8748764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8749126Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8749491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.8749939Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.8750326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.8750719Z return self.weight * hidden_states 2025-09-07T08:14:18.8750856Z 2025-09-07T08:14:18.8750950Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8751299Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8751620Z return mod(**inputs) 2025-09-07T08:14:18.8751948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8752306Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8752666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8753031Z layer_outputs = layer_module( 2025-09-07T08:14:18.8753372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8753735Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8754099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8754470Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8754835Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8755207Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8755560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.8755931Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.8756063Z 2025-09-07T08:14:18.8756157Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8756502Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8756819Z return mod(**inputs) 2025-09-07T08:14:18.8757141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8757498Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8757847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8758202Z layer_outputs = layer_module( 2025-09-07T08:14:18.8758527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8758877Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8759229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8759652Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8760059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8760419Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8760788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.8761148Z key_states = self.k(current_states) 2025-09-07T08:14:18.8761276Z 2025-09-07T08:14:18.8761377Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8761727Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8762036Z return mod(**inputs) 2025-09-07T08:14:18.8762372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8762737Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8763125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8763493Z layer_outputs = layer_module( 2025-09-07T08:14:18.8763827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8764214Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8764576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8764950Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8765303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8765669Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8766027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8766441Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8766622Z 2025-09-07T08:14:18.8766720Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8767055Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8767370Z return mod(**inputs) 2025-09-07T08:14:18.8767699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8768056Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8768398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8768757Z layer_outputs = layer_module( 2025-09-07T08:14:18.8769084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8769429Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8769790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8770155Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8770514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8770881Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8771251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8771656Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8771837Z 2025-09-07T08:14:18.8771929Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8772273Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8772579Z return mod(**inputs) 2025-09-07T08:14:18.8772962Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8773354Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8773697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8774065Z layer_outputs = layer_module( 2025-09-07T08:14:18.8774398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8774746Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8775106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8775467Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8775819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8776198Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8776583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8776991Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8777212Z 2025-09-07T08:14:18.8777338Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8777684Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8777988Z return mod(**inputs) 2025-09-07T08:14:18.8778323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8790225Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8790648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8791028Z layer_outputs = layer_module( 2025-09-07T08:14:18.8791377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8791743Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8792115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8792491Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8792854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8793219Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8793578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8794016Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8794221Z 2025-09-07T08:14:18.8794328Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8794685Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8795012Z return mod(**inputs) 2025-09-07T08:14:18.8795374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8795746Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8796097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8796456Z layer_outputs = layer_module( 2025-09-07T08:14:18.8796798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8797151Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8797510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8797874Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8798358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8798766Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8799126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8799562Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8799766Z 2025-09-07T08:14:18.8799866Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8800214Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8800526Z return mod(**inputs) 2025-09-07T08:14:18.8800855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8801216Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8801563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8801956Z layer_outputs = layer_module( 2025-09-07T08:14:18.8802287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8802678Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8803035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8803596Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8803955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8804314Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8804672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.8805045Z value_states = self.v(current_states) 2025-09-07T08:14:18.8805177Z 2025-09-07T08:14:18.8805276Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8805615Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8805927Z return mod(**inputs) 2025-09-07T08:14:18.8806255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8806608Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8806961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8807324Z layer_outputs = layer_module( 2025-09-07T08:14:18.8807662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8808013Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8808370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8808738Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8809091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8809456Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8809810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8810201Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8810360Z 2025-09-07T08:14:18.8810457Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8810800Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8811111Z return mod(**inputs) 2025-09-07T08:14:18.8811439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8811959Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8812310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8812662Z layer_outputs = layer_module( 2025-09-07T08:14:18.8812991Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8813338Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8813688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8814047Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8814397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8814751Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8815104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8815542Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8815694Z 2025-09-07T08:14:18.8815795Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8816178Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8816488Z return mod(**inputs) 2025-09-07T08:14:18.8816818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8817226Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8817577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8817930Z layer_outputs = layer_module( 2025-09-07T08:14:18.8818259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8818611Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8818972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8819332Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8819683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8820047Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8820400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8820784Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8820933Z 2025-09-07T08:14:18.8821030Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8821362Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8821671Z return mod(**inputs) 2025-09-07T08:14:18.8821998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8822354Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8822695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8823052Z layer_outputs = layer_module( 2025-09-07T08:14:18.8823377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8823722Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8824075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8824436Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8824790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8825238Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8825599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.8825985Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.8826141Z 2025-09-07T08:14:18.8826237Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8826577Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8826880Z return mod(**inputs) 2025-09-07T08:14:18.8827207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8827559Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8828478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8828836Z layer_outputs = layer_module( 2025-09-07T08:14:18.8829203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8829554Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8829939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8830302Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8830656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8831013Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8831362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.8831714Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.8831837Z 2025-09-07T08:14:18.8831909Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.8832127Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8832465Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8832776Z return mod(**inputs) 2025-09-07T08:14:18.8833102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8833169Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8833400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8833462Z layer_outputs = layer_module( 2025-09-07T08:14:18.8833680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8833749Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8833978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8834062Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8834293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.8834384Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.8834609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.8834682Z return self.weight * hidden_states 2025-09-07T08:14:18.8834685Z 2025-09-07T08:14:18.8834778Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8834967Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8835025Z return mod(**inputs) 2025-09-07T08:14:18.8835256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8835390Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8835656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8835722Z layer_outputs = layer_module( 2025-09-07T08:14:18.8835935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8836004Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8836234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8836311Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8836544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8836653Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8836882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.8836992Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.8836995Z 2025-09-07T08:14:18.8837085Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8837305Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8837363Z return mod(**inputs) 2025-09-07T08:14:18.8837602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8837667Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8837898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8837962Z layer_outputs = layer_module( 2025-09-07T08:14:18.8838172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8838244Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8838470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8838548Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8838777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8838876Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8839105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.8839179Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.8839182Z 2025-09-07T08:14:18.8839274Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8839457Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8839514Z return mod(**inputs) 2025-09-07T08:14:18.8839751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8839813Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8840059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8840122Z layer_outputs = layer_module( 2025-09-07T08:14:18.8840346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8840417Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8840652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8840732Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8840960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8841103Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8841366Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.8841437Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.8841441Z 2025-09-07T08:14:18.8841518Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.8841612Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8841810Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8841868Z return mod(**inputs) 2025-09-07T08:14:18.8842101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8842169Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8842398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8842462Z layer_outputs = layer_module( 2025-09-07T08:14:18.8842718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8842788Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8843057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8843130Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8843360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.8843459Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.8843688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.8843755Z return self.weight * hidden_states 2025-09-07T08:14:18.8843758Z 2025-09-07T08:14:18.8843851Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8844042Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8844097Z return mod(**inputs) 2025-09-07T08:14:18.8844331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8844394Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8844627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8844692Z layer_outputs = layer_module( 2025-09-07T08:14:18.8844905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8844973Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8845198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8845270Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8845502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8845576Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8845807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.8845874Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.8845878Z 2025-09-07T08:14:18.8845970Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8846153Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8846207Z return mod(**inputs) 2025-09-07T08:14:18.8846439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8846502Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8846809Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8846900Z layer_outputs = layer_module( 2025-09-07T08:14:18.8847111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8847185Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8847410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8847482Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8847706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8847776Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8848005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.8848074Z key_states = self.k(current_states) 2025-09-07T08:14:18.8848079Z 2025-09-07T08:14:18.8848206Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8848392Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8848452Z return mod(**inputs) 2025-09-07T08:14:18.8848712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8848776Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8849008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8849070Z layer_outputs = layer_module( 2025-09-07T08:14:18.8849288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8849354Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8849584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8849670Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8849895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8849968Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8850195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8850316Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8850322Z 2025-09-07T08:14:18.8850410Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8850592Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8850650Z return mod(**inputs) 2025-09-07T08:14:18.8850879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8850948Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8851178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8851240Z layer_outputs = layer_module( 2025-09-07T08:14:18.8851454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8851520Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8851748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8851816Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8852039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8852109Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8852375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8852524Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8852527Z 2025-09-07T08:14:18.8852617Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8852803Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8852859Z return mod(**inputs) 2025-09-07T08:14:18.8853088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8853151Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8853376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8853441Z layer_outputs = layer_module( 2025-09-07T08:14:18.8853652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8853755Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8853997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8854105Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8854341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8854412Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8854636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8854753Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8854756Z 2025-09-07T08:14:18.8854846Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8855034Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8855096Z return mod(**inputs) 2025-09-07T08:14:18.8855330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8855394Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8855623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8855686Z layer_outputs = layer_module( 2025-09-07T08:14:18.8855896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8855967Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8856194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8856264Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8856492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8856563Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8856790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8856936Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8856939Z 2025-09-07T08:14:18.8857033Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8857264Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8860319Z return mod(**inputs) 2025-09-07T08:14:18.8860597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8860671Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8860907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8861307Z layer_outputs = layer_module( 2025-09-07T08:14:18.8861529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8861608Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8861846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8861920Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8862152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8862249Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8862478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8862626Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8862632Z 2025-09-07T08:14:18.8862767Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8862971Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8863030Z return mod(**inputs) 2025-09-07T08:14:18.8863346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8863414Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8863645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8863714Z layer_outputs = layer_module( 2025-09-07T08:14:18.8863930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8864010Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8864238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8864317Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8864545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8864618Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8864852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.8864926Z value_states = self.v(current_states) 2025-09-07T08:14:18.8864930Z 2025-09-07T08:14:18.8865031Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8865229Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8865290Z return mod(**inputs) 2025-09-07T08:14:18.8865529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8865594Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8865832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8865896Z layer_outputs = layer_module( 2025-09-07T08:14:18.8866111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8866187Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8866418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8866497Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8866802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8866882Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8867108Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8867253Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8867256Z 2025-09-07T08:14:18.8867359Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8867553Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8867620Z return mod(**inputs) 2025-09-07T08:14:18.8867854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8867919Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8868160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8868224Z layer_outputs = layer_module( 2025-09-07T08:14:18.8868443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8868514Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8868779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8868857Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8869130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8869211Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8869437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8869537Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8869540Z 2025-09-07T08:14:18.8869631Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8869819Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8869878Z return mod(**inputs) 2025-09-07T08:14:18.8870111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8870177Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8870407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8870470Z layer_outputs = layer_module( 2025-09-07T08:14:18.8870687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8870755Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8870987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8871059Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8871289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8871360Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8871586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8871684Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8871688Z 2025-09-07T08:14:18.8871778Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8871960Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8872015Z return mod(**inputs) 2025-09-07T08:14:18.8872245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8872359Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8872590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8872655Z layer_outputs = layer_module( 2025-09-07T08:14:18.8872907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8872974Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8873210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8873280Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8873510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8873580Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8873809Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.8873903Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.8873907Z 2025-09-07T08:14:18.8873995Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8874211Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8874271Z return mod(**inputs) 2025-09-07T08:14:18.8874505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8874598Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8874830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8874895Z layer_outputs = layer_module( 2025-09-07T08:14:18.8875107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8875177Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8875403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8875471Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8875705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8875773Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8876002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.8876070Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.8876074Z 2025-09-07T08:14:18.8876164Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8876345Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8876402Z return mod(**inputs) 2025-09-07T08:14:18.8876633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8876693Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8876923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8876986Z layer_outputs = layer_module( 2025-09-07T08:14:18.8877196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8877265Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8877489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8877560Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8877784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:14:18.8877949Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:14:18.8877952Z 2025-09-07T08:14:18.8878024Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.8878111Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8878295Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8878393Z return mod(**inputs) 2025-09-07T08:14:18.8878625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8878689Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8878918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8878980Z layer_outputs = layer_module( 2025-09-07T08:14:18.8879190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8879259Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8879485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8879567Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8879828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.8879915Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.8880176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.8880245Z return self.weight * hidden_states 2025-09-07T08:14:18.8880248Z 2025-09-07T08:14:18.8880339Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8880522Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8880580Z return mod(**inputs) 2025-09-07T08:14:18.8880811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8880872Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8881105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8881168Z layer_outputs = layer_module( 2025-09-07T08:14:18.8881379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8881449Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8881675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8881755Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8881979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8882090Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8882318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.8882387Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.8882392Z 2025-09-07T08:14:18.8882483Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8882663Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8882723Z return mod(**inputs) 2025-09-07T08:14:18.8882951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8883012Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8883242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8883341Z layer_outputs = layer_module( 2025-09-07T08:14:18.8883553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8883618Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8883843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8883953Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8884180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8884286Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8884513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.8884585Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.8884592Z 2025-09-07T08:14:18.8884681Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8884861Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8884919Z return mod(**inputs) 2025-09-07T08:14:18.8885148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8885254Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8885487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8885583Z layer_outputs = layer_module( 2025-09-07T08:14:18.8885800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8885866Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8886095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8886171Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8886397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8886498Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8886726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.8886796Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.8886800Z 2025-09-07T08:14:18.8886869Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.8886961Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8887143Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8887198Z return mod(**inputs) 2025-09-07T08:14:18.8887430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8887494Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8887724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8887783Z layer_outputs = layer_module( 2025-09-07T08:14:18.8887996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8888066Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8888292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8888366Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8888593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.8888688Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.8888974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.8889041Z return self.weight * hidden_states 2025-09-07T08:14:18.8889044Z 2025-09-07T08:14:18.8889134Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8889317Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8889411Z return mod(**inputs) 2025-09-07T08:14:18.8889641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8889706Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8889935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8889996Z layer_outputs = layer_module( 2025-09-07T08:14:18.8890213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8890279Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8890504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8890578Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8890865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8890943Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8891203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.8891273Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.8891281Z 2025-09-07T08:14:18.8891368Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8891548Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8891608Z return mod(**inputs) 2025-09-07T08:14:18.8891837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8891902Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8892130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8892195Z layer_outputs = layer_module( 2025-09-07T08:14:18.8892408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8892477Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8892708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8892775Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8893002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8893077Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8893302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.8893374Z key_states = self.k(current_states) 2025-09-07T08:14:18.8893377Z 2025-09-07T08:14:18.8893468Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8893652Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8893708Z return mod(**inputs) 2025-09-07T08:14:18.8893937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8894002Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8894230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8894296Z layer_outputs = layer_module( 2025-09-07T08:14:18.8894565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8894634Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8894865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8894970Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8895201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8895272Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8895498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8895623Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8895627Z 2025-09-07T08:14:18.8895716Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8895905Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8895960Z return mod(**inputs) 2025-09-07T08:14:18.8896191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8896256Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8896516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8896582Z layer_outputs = layer_module( 2025-09-07T08:14:18.8896827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8896896Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8897123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8897251Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8897487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8897556Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8897784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8897896Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8897899Z 2025-09-07T08:14:18.8897987Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8898172Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8898227Z return mod(**inputs) 2025-09-07T08:14:18.8898457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8898518Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8898751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8898810Z layer_outputs = layer_module( 2025-09-07T08:14:18.8899020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8899090Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8899318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8899386Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8899612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8899682Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8899906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8900062Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8900065Z 2025-09-07T08:14:18.8900155Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8900335Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8900394Z return mod(**inputs) 2025-09-07T08:14:18.8900665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8900730Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8900963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8901024Z layer_outputs = layer_module( 2025-09-07T08:14:18.8901238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8901305Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8901532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8901608Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8901833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8901941Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8902169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8902351Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8902358Z 2025-09-07T08:14:18.8902449Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8902634Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8902692Z return mod(**inputs) 2025-09-07T08:14:18.8902921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8902988Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8903443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8903505Z layer_outputs = layer_module( 2025-09-07T08:14:18.8903723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8903790Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8904024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8904095Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8904321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8904398Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8904624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8904767Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8904771Z 2025-09-07T08:14:18.8904862Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8905053Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8905110Z return mod(**inputs) 2025-09-07T08:14:18.8905341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8905405Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8905634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8905698Z layer_outputs = layer_module( 2025-09-07T08:14:18.8906000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8906071Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8906320Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8906439Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8906675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8906747Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8906975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.8907052Z value_states = self.v(current_states) 2025-09-07T08:14:18.8907056Z 2025-09-07T08:14:18.8907151Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8907345Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8907405Z return mod(**inputs) 2025-09-07T08:14:18.8907645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8907711Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8907988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8908056Z layer_outputs = layer_module( 2025-09-07T08:14:18.8908317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8908396Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8908626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8908697Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8908932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8909004Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8909237Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8909338Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8909342Z 2025-09-07T08:14:18.8909440Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8909640Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8909700Z return mod(**inputs) 2025-09-07T08:14:18.8909935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8910001Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8910236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8910303Z layer_outputs = layer_module( 2025-09-07T08:14:18.8910517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8910590Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8910825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8910903Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8911131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8911203Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8911436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8911531Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8911571Z 2025-09-07T08:14:18.8911665Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8911849Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8911911Z return mod(**inputs) 2025-09-07T08:14:18.8912145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8912259Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8912491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8912555Z layer_outputs = layer_module( 2025-09-07T08:14:18.8912773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8912844Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8913070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8913147Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8913375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8913447Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8913710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8913804Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8913811Z 2025-09-07T08:14:18.8913931Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8914117Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8914181Z return mod(**inputs) 2025-09-07T08:14:18.8914413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8914485Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8914714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8914778Z layer_outputs = layer_module( 2025-09-07T08:14:18.8914996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8915063Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8915296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8915369Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8915594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8915669Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8915896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.8915992Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.8915996Z 2025-09-07T08:14:18.8916084Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8916270Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8916335Z return mod(**inputs) 2025-09-07T08:14:18.8916563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8916630Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8916861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8916929Z layer_outputs = layer_module( 2025-09-07T08:14:18.8917139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8917243Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8917477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8917546Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8917781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8917882Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8918113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.8918189Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.8918193Z 2025-09-07T08:14:18.8918266Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.8918373Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8918555Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8918614Z return mod(**inputs) 2025-09-07T08:14:18.8918849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8918913Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8919181Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8919244Z layer_outputs = layer_module( 2025-09-07T08:14:18.8919463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8919561Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8919792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8919877Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8920106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.8920202Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.8920430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.8920498Z return self.weight * hidden_states 2025-09-07T08:14:18.8920503Z 2025-09-07T08:14:18.8920603Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8920786Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8920847Z return mod(**inputs) 2025-09-07T08:14:18.8921078Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8921142Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8921378Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8921443Z layer_outputs = layer_module( 2025-09-07T08:14:18.8921659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8921727Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8921959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8922039Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8922265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8922383Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8922608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.8922683Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.8922686Z 2025-09-07T08:14:18.8922812Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8922994Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8923059Z return mod(**inputs) 2025-09-07T08:14:18.8923290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8923391Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8923623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8923691Z layer_outputs = layer_module( 2025-09-07T08:14:18.8923902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8923970Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8924207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8924285Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8924516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8924614Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8924872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.8924953Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.8924955Z 2025-09-07T08:14:18.8925045Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8925261Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8925319Z return mod(**inputs) 2025-09-07T08:14:18.8925549Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8925618Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8925851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8925917Z layer_outputs = layer_module( 2025-09-07T08:14:18.8926135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8926218Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8926454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8926536Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8926770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8926879Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8927110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.8927183Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.8927186Z 2025-09-07T08:14:18.8927259Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.8927362Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8927557Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8927623Z return mod(**inputs) 2025-09-07T08:14:18.8927857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8927926Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8928163Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8928225Z layer_outputs = layer_module( 2025-09-07T08:14:18.8928448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8928556Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8928791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8928862Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8929093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.8929230Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.8929460Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.8929531Z return self.weight * hidden_states 2025-09-07T08:14:18.8929535Z 2025-09-07T08:14:18.8929629Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8929815Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8929880Z return mod(**inputs) 2025-09-07T08:14:18.8930109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8930177Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8930408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8930508Z layer_outputs = layer_module( 2025-09-07T08:14:18.8930732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8930830Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8931064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8931134Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8931364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8931438Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8931662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.8931732Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.8931735Z 2025-09-07T08:14:18.8931831Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8932018Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8932073Z return mod(**inputs) 2025-09-07T08:14:18.8932304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8932367Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8932597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8932662Z layer_outputs = layer_module( 2025-09-07T08:14:18.8932875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8932944Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8933171Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8933244Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8933478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8933551Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8933779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.8933845Z key_states = self.k(current_states) 2025-09-07T08:14:18.8933848Z 2025-09-07T08:14:18.8933937Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8934188Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8934244Z return mod(**inputs) 2025-09-07T08:14:18.8934475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8934536Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8934803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8934868Z layer_outputs = layer_module( 2025-09-07T08:14:18.8935081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8935148Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8935373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8935442Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8935669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8935738Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8935964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8936123Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8936127Z 2025-09-07T08:14:18.8936218Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8936431Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8936487Z return mod(**inputs) 2025-09-07T08:14:18.8936718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8936779Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8937011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8937073Z layer_outputs = layer_module( 2025-09-07T08:14:18.8937326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8937394Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8937624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8937695Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8937920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8937991Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8938215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8938325Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8938330Z 2025-09-07T08:14:18.8938421Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8938602Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8938660Z return mod(**inputs) 2025-09-07T08:14:18.8938891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8938952Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8939186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8939247Z layer_outputs = layer_module( 2025-09-07T08:14:18.8939460Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8939524Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8939808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8939879Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8940104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8940213Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8940441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8940553Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8940558Z 2025-09-07T08:14:18.8940648Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8940829Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8940887Z return mod(**inputs) 2025-09-07T08:14:18.8941115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8941184Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8941412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8941473Z layer_outputs = layer_module( 2025-09-07T08:14:18.8941730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8941801Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8942063Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8942132Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8942361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8942430Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8942655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8942803Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8942807Z 2025-09-07T08:14:18.8942897Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8943085Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8943139Z return mod(**inputs) 2025-09-07T08:14:18.8943369Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8943433Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8943659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8943722Z layer_outputs = layer_module( 2025-09-07T08:14:18.8943933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8943999Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8944227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8944298Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8944529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8944596Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8944825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8944959Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8944962Z 2025-09-07T08:14:18.8945048Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8945232Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8945345Z return mod(**inputs) 2025-09-07T08:14:18.8945578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8945639Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8945912Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8945976Z layer_outputs = layer_module( 2025-09-07T08:14:18.8946190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8946258Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8946482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8946557Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8946784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8946851Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8947079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.8947149Z value_states = self.v(current_states) 2025-09-07T08:14:18.8947185Z 2025-09-07T08:14:18.8947279Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8947460Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8947546Z return mod(**inputs) 2025-09-07T08:14:18.8947786Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8947849Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8948081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8948142Z layer_outputs = layer_module( 2025-09-07T08:14:18.8948357Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8948427Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8948654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8948728Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8948954Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8949027Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8949253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8949353Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8949358Z 2025-09-07T08:14:18.8949454Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8949635Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8949695Z return mod(**inputs) 2025-09-07T08:14:18.8949925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8949988Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8950215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8950278Z layer_outputs = layer_module( 2025-09-07T08:14:18.8950492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8950557Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8950784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8950893Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8951121Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8951194Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8951458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8951552Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8951555Z 2025-09-07T08:14:18.8951653Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8951842Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8951902Z return mod(**inputs) 2025-09-07T08:14:18.8952143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8952212Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8952441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8952504Z layer_outputs = layer_module( 2025-09-07T08:14:18.8952723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8952857Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8953094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8953214Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8953446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8953519Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8953749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8953857Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8953861Z 2025-09-07T08:14:18.8953957Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8954148Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8954211Z return mod(**inputs) 2025-09-07T08:14:18.8954446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8954518Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8954750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8954832Z layer_outputs = layer_module( 2025-09-07T08:14:18.8955047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8955120Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8955354Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8955427Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8955661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8955736Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8955970Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.8956067Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.8956070Z 2025-09-07T08:14:18.8956170Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8956363Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8956423Z return mod(**inputs) 2025-09-07T08:14:18.8956720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8956784Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8957017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8957129Z layer_outputs = layer_module( 2025-09-07T08:14:18.8957342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8957418Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8957648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8957730Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8957958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8958032Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8958264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.8958334Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.8958337Z 2025-09-07T08:14:18.8958434Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8958653Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8958715Z return mod(**inputs) 2025-09-07T08:14:18.8958990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8959055Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8959292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8959356Z layer_outputs = layer_module( 2025-09-07T08:14:18.8959571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8959651Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8959880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8959961Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8960188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:14:18.8960325Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:14:18.8960328Z 2025-09-07T08:14:18.8960403Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.8960493Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8960683Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8960741Z return mod(**inputs) 2025-09-07T08:14:18.8960976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8961037Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8961265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8961337Z layer_outputs = layer_module( 2025-09-07T08:14:18.8961548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8961622Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8961853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8961938Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8962171Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.8962295Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.8962529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.8962599Z return self.weight * hidden_states 2025-09-07T08:14:18.8962602Z 2025-09-07T08:14:18.8962735Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8962918Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8962981Z return mod(**inputs) 2025-09-07T08:14:18.8963219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8963282Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8963514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8963577Z layer_outputs = layer_module( 2025-09-07T08:14:18.8963791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8963865Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8964093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8964214Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8964444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8964594Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8964832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.8964903Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.8964907Z 2025-09-07T08:14:18.8965005Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8965192Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8965255Z return mod(**inputs) 2025-09-07T08:14:18.8965488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8965553Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8965794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8965857Z layer_outputs = layer_module( 2025-09-07T08:14:18.8966082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8966148Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8966375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8966459Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8966686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8966791Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8967018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.8967099Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.8967102Z 2025-09-07T08:14:18.8967192Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8967375Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8967444Z return mod(**inputs) 2025-09-07T08:14:18.8967676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8967749Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8968017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8968080Z layer_outputs = layer_module( 2025-09-07T08:14:18.8968296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8968363Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8968636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.8968714Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.8968946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.8969057Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.8969283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.8969357Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.8969360Z 2025-09-07T08:14:18.8969431Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.8969526Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8969705Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8969765Z return mod(**inputs) 2025-09-07T08:14:18.8970033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8970099Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8970374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8970439Z layer_outputs = layer_module( 2025-09-07T08:14:18.8970652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8970729Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8970953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8971033Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8971262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.8971366Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.8971602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.8971670Z return self.weight * hidden_states 2025-09-07T08:14:18.8971673Z 2025-09-07T08:14:18.8971765Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8971947Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8972004Z return mod(**inputs) 2025-09-07T08:14:18.8972233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8972295Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8972525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8972587Z layer_outputs = layer_module( 2025-09-07T08:14:18.8972802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8972869Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8973096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8973172Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8973399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8973515Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8973743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.8973813Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.8973821Z 2025-09-07T08:14:18.8973913Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8974132Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8974201Z return mod(**inputs) 2025-09-07T08:14:18.8974430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8974497Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8974727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8974789Z layer_outputs = layer_module( 2025-09-07T08:14:18.8975007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8975072Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8975307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8975377Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8975654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8975731Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8975990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.8976062Z key_states = self.k(current_states) 2025-09-07T08:14:18.8976065Z 2025-09-07T08:14:18.8976155Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8976343Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8976402Z return mod(**inputs) 2025-09-07T08:14:18.8976632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8976703Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8976934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8977002Z layer_outputs = layer_module( 2025-09-07T08:14:18.8977270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8977338Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8977569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8977640Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8977876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8977947Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8978172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8978295Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8978301Z 2025-09-07T08:14:18.8978390Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8978573Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8978629Z return mod(**inputs) 2025-09-07T08:14:18.8978861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8978922Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8979150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8979250Z layer_outputs = layer_module( 2025-09-07T08:14:18.8979461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8979529Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8979755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8979860Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8980092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8980163Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8980392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8980505Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8980511Z 2025-09-07T08:14:18.8980614Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8980801Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8980859Z return mod(**inputs) 2025-09-07T08:14:18.8981127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8981194Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8981428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8981521Z layer_outputs = layer_module( 2025-09-07T08:14:18.8981739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8981813Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8982042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8982120Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8982349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8982421Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8982662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.8982775Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.8982778Z 2025-09-07T08:14:18.8982876Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8983055Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8983112Z return mod(**inputs) 2025-09-07T08:14:18.8983342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8983407Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8983640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8983700Z layer_outputs = layer_module( 2025-09-07T08:14:18.8983921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8983991Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8984216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8984294Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8984522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8984598Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8984826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8985009Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8985016Z 2025-09-07T08:14:18.8985105Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8985288Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8985386Z return mod(**inputs) 2025-09-07T08:14:18.8985617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8985684Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8985915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8985974Z layer_outputs = layer_module( 2025-09-07T08:14:18.8986189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8986259Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8986497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8986566Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8986824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8986901Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8987160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.8987306Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.8987310Z 2025-09-07T08:14:18.8987398Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8987589Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8987650Z return mod(**inputs) 2025-09-07T08:14:18.8987883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8987953Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8988183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8988249Z layer_outputs = layer_module( 2025-09-07T08:14:18.8988461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8988530Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8988763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8988832Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8989066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8989138Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8989367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.8989447Z value_states = self.v(current_states) 2025-09-07T08:14:18.8989452Z 2025-09-07T08:14:18.8989545Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8989734Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8989793Z return mod(**inputs) 2025-09-07T08:14:18.8990030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8990095Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8990324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8990423Z layer_outputs = layer_module( 2025-09-07T08:14:18.8990634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8990707Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8990933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8991039Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8991277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8991354Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8991588Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8991691Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8991694Z 2025-09-07T08:14:18.8991795Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8991984Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8992044Z return mod(**inputs) 2025-09-07T08:14:18.8992281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8992383Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8992624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8992688Z layer_outputs = layer_module( 2025-09-07T08:14:18.8992932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8993007Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8993236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8993314Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8993540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8993609Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8993839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8993932Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8993936Z 2025-09-07T08:14:18.8994028Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8994212Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8994269Z return mod(**inputs) 2025-09-07T08:14:18.8994497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8994559Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8994793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8994855Z layer_outputs = layer_module( 2025-09-07T08:14:18.8995070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8995140Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8995366Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8995443Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8995673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8995747Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8995976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.8996124Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.8996132Z 2025-09-07T08:14:18.8996225Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8996409Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8996472Z return mod(**inputs) 2025-09-07T08:14:18.8996762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8996836Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8997068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8997131Z layer_outputs = layer_module( 2025-09-07T08:14:18.8997350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8997415Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8997650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8997720Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.8997945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.8998078Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.8998306Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.8998432Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.8998436Z 2025-09-07T08:14:18.8998525Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.8998713Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.8998770Z return mod(**inputs) 2025-09-07T08:14:18.8998999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.8999069Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.8999299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.8999369Z layer_outputs = layer_module( 2025-09-07T08:14:18.8999586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.8999654Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.8999891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.8999962Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9000197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9000268Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9000496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9000572Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9000575Z 2025-09-07T08:14:18.9000647Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9000745Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9000931Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9000988Z return mod(**inputs) 2025-09-07T08:14:18.9001228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9001294Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9001532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9001595Z layer_outputs = layer_module( 2025-09-07T08:14:18.9001848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9001916Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9002145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9002267Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9002497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.9002595Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9002822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9002894Z return self.weight * hidden_states 2025-09-07T08:14:18.9002897Z 2025-09-07T08:14:18.9002997Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9003343Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9003403Z return mod(**inputs) 2025-09-07T08:14:18.9003633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9003698Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9004024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9004090Z layer_outputs = layer_module( 2025-09-07T08:14:18.9004368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9004437Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9004668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9004747Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9004974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9005089Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9005318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.9005398Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.9005402Z 2025-09-07T08:14:18.9005495Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9005685Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9005753Z return mod(**inputs) 2025-09-07T08:14:18.9005986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9006061Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9006291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9006359Z layer_outputs = layer_module( 2025-09-07T08:14:18.9006573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9006641Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9006877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9006955Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9007193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9007296Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9007525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.9007660Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.9007663Z 2025-09-07T08:14:18.9007754Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9007944Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9008002Z return mod(**inputs) 2025-09-07T08:14:18.9008287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9008358Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9008592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9008662Z layer_outputs = layer_module( 2025-09-07T08:14:18.9008875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9008945Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9009173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9009252Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9009488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9009621Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9009863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.9009934Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.9009966Z 2025-09-07T08:14:18.9010042Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9010139Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9010322Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9010388Z return mod(**inputs) 2025-09-07T08:14:18.9010621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9010685Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9010924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9010987Z layer_outputs = layer_module( 2025-09-07T08:14:18.9011209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9011276Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9011515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9011589Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9011815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.9011926Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9012154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9012229Z return self.weight * hidden_states 2025-09-07T08:14:18.9012232Z 2025-09-07T08:14:18.9012319Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9012505Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9012567Z return mod(**inputs) 2025-09-07T08:14:18.9012800Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9012868Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9013097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9013168Z layer_outputs = layer_module( 2025-09-07T08:14:18.9019288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9019371Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9019644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9019805Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9020065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9020146Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9020393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9020469Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9020474Z 2025-09-07T08:14:18.9020581Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9020796Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9020862Z return mod(**inputs) 2025-09-07T08:14:18.9021111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9021184Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9021455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9021528Z layer_outputs = layer_module( 2025-09-07T08:14:18.9021786Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9021861Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9022102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9022176Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9022409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9022487Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9022720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9022788Z key_states = self.k(current_states) 2025-09-07T08:14:18.9022794Z 2025-09-07T08:14:18.9022894Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9023093Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9023154Z return mod(**inputs) 2025-09-07T08:14:18.9023390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9023455Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9023687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9023756Z layer_outputs = layer_module( 2025-09-07T08:14:18.9023970Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9024045Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9024272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9024347Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9024596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9024675Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9024915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9025042Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9025099Z 2025-09-07T08:14:18.9025205Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9025404Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9025464Z return mod(**inputs) 2025-09-07T08:14:18.9025710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9025814Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9026053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9026116Z layer_outputs = layer_module( 2025-09-07T08:14:18.9026335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9026412Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9026645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9026727Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9026954Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9027033Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9027295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9027411Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9027415Z 2025-09-07T08:14:18.9027538Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9027734Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9027797Z return mod(**inputs) 2025-09-07T08:14:18.9028028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9028096Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9028331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9028394Z layer_outputs = layer_module( 2025-09-07T08:14:18.9028612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9028682Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9028908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9028987Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9029214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9029295Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9029520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9029639Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9029642Z 2025-09-07T08:14:18.9029735Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9029921Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9029989Z return mod(**inputs) 2025-09-07T08:14:18.9030220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9030287Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9030520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9030583Z layer_outputs = layer_module( 2025-09-07T08:14:18.9030802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9030906Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9031140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9031211Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9031445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9031551Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9031777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9031933Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9031937Z 2025-09-07T08:14:18.9032029Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9032218Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9032277Z return mod(**inputs) 2025-09-07T08:14:18.9032520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9032590Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9032823Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9032925Z layer_outputs = layer_module( 2025-09-07T08:14:18.9033143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9033256Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9033503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9033583Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9033818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9033897Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9034132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9034279Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9034284Z 2025-09-07T08:14:18.9034385Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9034586Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9034644Z return mod(**inputs) 2025-09-07T08:14:18.9034886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9034951Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9035182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9035251Z layer_outputs = layer_module( 2025-09-07T08:14:18.9035465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9035537Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9035767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9035841Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9036066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9036139Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9036366Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9036434Z value_states = self.v(current_states) 2025-09-07T08:14:18.9036438Z 2025-09-07T08:14:18.9036530Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9036754Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9036810Z return mod(**inputs) 2025-09-07T08:14:18.9037042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9037142Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9037379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9037439Z layer_outputs = layer_module( 2025-09-07T08:14:18.9037654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9037725Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9037951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9038023Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9038247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9038321Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9038579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9038682Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9038685Z 2025-09-07T08:14:18.9038783Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9039001Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9039062Z return mod(**inputs) 2025-09-07T08:14:18.9039296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9039359Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9039596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9039658Z layer_outputs = layer_module( 2025-09-07T08:14:18.9039873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9039942Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9040170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9040244Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9040472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9040547Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9040804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9040909Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9040912Z 2025-09-07T08:14:18.9041004Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9041188Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9041248Z return mod(**inputs) 2025-09-07T08:14:18.9041479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9041542Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9041772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9041831Z layer_outputs = layer_module( 2025-09-07T08:14:18.9042043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9042108Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9042379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9042447Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9042673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9042783Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9043010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9043106Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9043110Z 2025-09-07T08:14:18.9043198Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9043383Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9043438Z return mod(**inputs) 2025-09-07T08:14:18.9043667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9043731Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9043960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9044025Z layer_outputs = layer_module( 2025-09-07T08:14:18.9044272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9044337Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9044612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9044682Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9044916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9044986Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9045214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9045310Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9045313Z 2025-09-07T08:14:18.9045403Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9045593Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9045648Z return mod(**inputs) 2025-09-07T08:14:18.9045878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9045938Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9046167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9046228Z layer_outputs = layer_module( 2025-09-07T08:14:18.9046442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9046515Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9046742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9046810Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9047046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9047115Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9047354Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9047421Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9047424Z 2025-09-07T08:14:18.9047515Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9047697Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9047788Z return mod(**inputs) 2025-09-07T08:14:18.9048025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9048087Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9048321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9048443Z layer_outputs = layer_module( 2025-09-07T08:14:18.9048655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9048723Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9048949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9049019Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9049251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:14:18.9049377Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:14:18.9049385Z 2025-09-07T08:14:18.9049458Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9049549Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9049772Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9049830Z return mod(**inputs) 2025-09-07T08:14:18.9050099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9050162Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9050391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9050452Z layer_outputs = layer_module( 2025-09-07T08:14:18.9050662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9050730Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9050955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9051037Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9051267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.9051352Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9051580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9051650Z return self.weight * hidden_states 2025-09-07T08:14:18.9051654Z 2025-09-07T08:14:18.9051750Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9051939Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9051998Z return mod(**inputs) 2025-09-07T08:14:18.9052231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9052293Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9052526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9052589Z layer_outputs = layer_module( 2025-09-07T08:14:18.9052806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9052876Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9053102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9053184Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9053409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9053559Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9053790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.9053897Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.9053902Z 2025-09-07T08:14:18.9054004Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9054193Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9054260Z return mod(**inputs) 2025-09-07T08:14:18.9054492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9054557Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9054794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9054868Z layer_outputs = layer_module( 2025-09-07T08:14:18.9055089Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9055158Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9055422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9055509Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9055770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9055879Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9056106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.9056180Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.9056191Z 2025-09-07T08:14:18.9056283Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9056468Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9056532Z return mod(**inputs) 2025-09-07T08:14:18.9056763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9056835Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9057065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9057131Z layer_outputs = layer_module( 2025-09-07T08:14:18.9057409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9057482Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9057725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9057811Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9058040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9058146Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9058374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.9058445Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.9058448Z 2025-09-07T08:14:18.9058521Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9058614Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9058799Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9058855Z return mod(**inputs) 2025-09-07T08:14:18.9059087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9059188Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9059419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9059481Z layer_outputs = layer_module( 2025-09-07T08:14:18.9059733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9059804Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9060045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9060116Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9060342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.9060438Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9060665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9060732Z return self.weight * hidden_states 2025-09-07T08:14:18.9060735Z 2025-09-07T08:14:18.9060826Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9061040Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9061099Z return mod(**inputs) 2025-09-07T08:14:18.9061330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9061422Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9061665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9061729Z layer_outputs = layer_module( 2025-09-07T08:14:18.9061945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9062012Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9062236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9062308Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9062534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9062608Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9062836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9062904Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9062911Z 2025-09-07T08:14:18.9062999Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9063179Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9063238Z return mod(**inputs) 2025-09-07T08:14:18.9063465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9063529Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9063757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9063819Z layer_outputs = layer_module( 2025-09-07T08:14:18.9064031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9064097Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9064324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9064391Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9064621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9064728Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9064952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9065021Z key_states = self.k(current_states) 2025-09-07T08:14:18.9065063Z 2025-09-07T08:14:18.9065152Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9065334Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9065391Z return mod(**inputs) 2025-09-07T08:14:18.9065620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9065685Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9065913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9065975Z layer_outputs = layer_module( 2025-09-07T08:14:18.9066185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9066249Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9066478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9066595Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9066835Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9066942Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9067180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9067309Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9067313Z 2025-09-07T08:14:18.9067405Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9067599Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9067655Z return mod(**inputs) 2025-09-07T08:14:18.9067892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9067958Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9068191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9068255Z layer_outputs = layer_module( 2025-09-07T08:14:18.9068468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9068540Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9068765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9068837Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9069070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9069142Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9069378Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9069494Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9069497Z 2025-09-07T08:14:18.9069586Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9069777Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9069832Z return mod(**inputs) 2025-09-07T08:14:18.9070065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9070128Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9070407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9070471Z layer_outputs = layer_module( 2025-09-07T08:14:18.9070682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9070783Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9071012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9071088Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9071315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9071385Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9071620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9071733Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9071736Z 2025-09-07T08:14:18.9071831Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9072013Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9072071Z return mod(**inputs) 2025-09-07T08:14:18.9072332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9072394Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9072657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9072718Z layer_outputs = layer_module( 2025-09-07T08:14:18.9072933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9072998Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9073225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9073294Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9073519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9073592Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9073817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9073965Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9073969Z 2025-09-07T08:14:18.9074056Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9074237Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9074295Z return mod(**inputs) 2025-09-07T08:14:18.9074525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9074588Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9074816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9074876Z layer_outputs = layer_module( 2025-09-07T08:14:18.9075092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9075156Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9075385Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9075451Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9075680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9075782Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9076009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9076146Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9076150Z 2025-09-07T08:14:18.9076272Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9076456Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9076510Z return mod(**inputs) 2025-09-07T08:14:18.9076740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9076804Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9077030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9077095Z layer_outputs = layer_module( 2025-09-07T08:14:18.9077307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9077370Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9077598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9077700Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9077934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9078005Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9078267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9078336Z value_states = self.v(current_states) 2025-09-07T08:14:18.9078340Z 2025-09-07T08:14:18.9078425Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9078612Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9078667Z return mod(**inputs) 2025-09-07T08:14:18.9078898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9078959Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9079187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9079252Z layer_outputs = layer_module( 2025-09-07T08:14:18.9079466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9079536Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9079759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9079827Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9080060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9080129Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9080361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9080464Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9080468Z 2025-09-07T08:14:18.9080559Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9080743Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9080799Z return mod(**inputs) 2025-09-07T08:14:18.9081032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9081094Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9081327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9081421Z layer_outputs = layer_module( 2025-09-07T08:14:18.9081633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9081703Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9081965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9082036Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9082262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9082335Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9082559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9082651Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9082656Z 2025-09-07T08:14:18.9082751Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9082942Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9083003Z return mod(**inputs) 2025-09-07T08:14:18.9083534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9083600Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9083866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9083926Z layer_outputs = layer_module( 2025-09-07T08:14:18.9084141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9084207Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9084433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9084511Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9084735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9084808Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9085037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9085133Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9085136Z 2025-09-07T08:14:18.9085225Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9085407Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9085466Z return mod(**inputs) 2025-09-07T08:14:18.9085696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9085762Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9085993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9086054Z layer_outputs = layer_module( 2025-09-07T08:14:18.9086274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9086341Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9086571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9086641Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9086869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9086944Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9087169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9087296Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9087300Z 2025-09-07T08:14:18.9087386Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9087568Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9087672Z return mod(**inputs) 2025-09-07T08:14:18.9087900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9087965Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9088195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9088259Z layer_outputs = layer_module( 2025-09-07T08:14:18.9088470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9088540Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9088772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9088838Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9089098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9089169Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9089429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9089498Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9089502Z 2025-09-07T08:14:18.9089572Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9089671Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9089859Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9089922Z return mod(**inputs) 2025-09-07T08:14:18.9090156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9090222Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9090459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9090522Z layer_outputs = layer_module( 2025-09-07T08:14:18.9090742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9090815Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9091050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9091135Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9091363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.9091454Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9091680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9091753Z return self.weight * hidden_states 2025-09-07T08:14:18.9091758Z 2025-09-07T08:14:18.9091848Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9092033Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9092093Z return mod(**inputs) 2025-09-07T08:14:18.9092325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9092388Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9092615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9092711Z layer_outputs = layer_module( 2025-09-07T08:14:18.9092930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9092995Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9093230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9093342Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9093573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9093691Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9093920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.9093998Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.9094003Z 2025-09-07T08:14:18.9094091Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9094279Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9094335Z return mod(**inputs) 2025-09-07T08:14:18.9094567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9094669Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9094903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9095001Z layer_outputs = layer_module( 2025-09-07T08:14:18.9095217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9095284Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9095518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9095598Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9095833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9095936Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9096166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.9096241Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.9096244Z 2025-09-07T08:14:18.9096333Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9096521Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9096580Z return mod(**inputs) 2025-09-07T08:14:18.9096813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9096878Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9097105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9097229Z layer_outputs = layer_module( 2025-09-07T08:14:18.9097443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9097520Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9097746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9097822Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9098053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9098150Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9098378Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.9098484Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.9098487Z 2025-09-07T08:14:18.9098565Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9098654Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9098838Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9098933Z return mod(**inputs) 2025-09-07T08:14:18.9099164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9099232Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9099462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9099522Z layer_outputs = layer_module( 2025-09-07T08:14:18.9099736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9099803Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9100033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9100102Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9100362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.9100462Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9100719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9100787Z return self.weight * hidden_states 2025-09-07T08:14:18.9100790Z 2025-09-07T08:14:18.9100877Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9101060Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9101118Z return mod(**inputs) 2025-09-07T08:14:18.9101346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9101411Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9101641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9101708Z layer_outputs = layer_module( 2025-09-07T08:14:18.9101919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9101984Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9102218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9102291Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9102536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9102614Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9102840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9102916Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9102920Z 2025-09-07T08:14:18.9103010Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9103365Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9103421Z return mod(**inputs) 2025-09-07T08:14:18.9103655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9103716Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9103942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9104074Z layer_outputs = layer_module( 2025-09-07T08:14:18.9104284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9104352Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9104578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9104697Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9104925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9104996Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9105229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9105295Z key_states = self.k(current_states) 2025-09-07T08:14:18.9105299Z 2025-09-07T08:14:18.9105394Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9105579Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9105636Z return mod(**inputs) 2025-09-07T08:14:18.9105873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9105935Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9106215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9106279Z layer_outputs = layer_module( 2025-09-07T08:14:18.9106536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9106611Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9106838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9106913Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9107136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9107206Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9107439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9107560Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9107563Z 2025-09-07T08:14:18.9107653Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9107836Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9107896Z return mod(**inputs) 2025-09-07T08:14:18.9108126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9108187Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9108420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9108481Z layer_outputs = layer_module( 2025-09-07T08:14:18.9108692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9108759Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9108986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9109059Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9109286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9109358Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9109586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9109746Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9109751Z 2025-09-07T08:14:18.9109841Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9110022Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9110082Z return mod(**inputs) 2025-09-07T08:14:18.9110350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9110418Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9110651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9110714Z layer_outputs = layer_module( 2025-09-07T08:14:18.9110930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9110998Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9111231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9111299Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9111525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9111634Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9111861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9112008Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9112011Z 2025-09-07T08:14:18.9112101Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9112285Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9112341Z return mod(**inputs) 2025-09-07T08:14:18.9112571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9112640Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9112868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9112933Z layer_outputs = layer_module( 2025-09-07T08:14:18.9113146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9113214Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9113444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9113513Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9113742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9113811Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9114037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9114184Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9114187Z 2025-09-07T08:14:18.9114274Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9114463Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9114521Z return mod(**inputs) 2025-09-07T08:14:18.9114751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9114812Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9115039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9115101Z layer_outputs = layer_module( 2025-09-07T08:14:18.9115351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9115422Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9115653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9115761Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9115996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9116067Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9116302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9116444Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9116448Z 2025-09-07T08:14:18.9116541Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9116731Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9116789Z return mod(**inputs) 2025-09-07T08:14:18.9117026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9117088Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9117355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9117419Z layer_outputs = layer_module( 2025-09-07T08:14:18.9117677Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9117748Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9117985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9118058Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9118285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9118355Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9118581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9118653Z value_states = self.v(current_states) 2025-09-07T08:14:18.9118656Z 2025-09-07T08:14:18.9118749Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9118932Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9118990Z return mod(**inputs) 2025-09-07T08:14:18.9119218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9119280Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9119512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9119573Z layer_outputs = layer_module( 2025-09-07T08:14:18.9119787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9119851Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9120080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9120153Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9120379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9120452Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9120676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9120777Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9120818Z 2025-09-07T08:14:18.9120910Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9121093Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9121153Z return mod(**inputs) 2025-09-07T08:14:18.9121382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9121479Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9121709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9121772Z layer_outputs = layer_module( 2025-09-07T08:14:18.9121984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9122050Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9122279Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9122351Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9122579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9122650Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9122909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9123005Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9123008Z 2025-09-07T08:14:18.9123124Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9123311Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9123366Z return mod(**inputs) 2025-09-07T08:14:18.9123597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9123663Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9123890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9123952Z layer_outputs = layer_module( 2025-09-07T08:14:18.9124165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9124231Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9124458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9124528Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9124754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9124822Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9125047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9125148Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9125151Z 2025-09-07T08:14:18.9125241Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9125429Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9125489Z return mod(**inputs) 2025-09-07T08:14:18.9125727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9125792Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9126021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9126090Z layer_outputs = layer_module( 2025-09-07T08:14:18.9126302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9126409Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9126638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9126708Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9126964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9127068Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9127305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9127398Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9127401Z 2025-09-07T08:14:18.9127492Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9127684Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9127742Z return mod(**inputs) 2025-09-07T08:14:18.9127983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9128048Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9128285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9128382Z layer_outputs = layer_module( 2025-09-07T08:14:18.9128597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9128717Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9128951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9129027Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9129253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9129327Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9129558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9129628Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9129632Z 2025-09-07T08:14:18.9129730Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9129914Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9129972Z return mod(**inputs) 2025-09-07T08:14:18.9130211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9130275Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9130509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9130570Z layer_outputs = layer_module( 2025-09-07T08:14:18.9130790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9130858Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9131087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9131165Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9131392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:14:18.9131523Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:14:18.9131527Z 2025-09-07T08:14:18.9131600Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9131692Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9131889Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9131980Z return mod(**inputs) 2025-09-07T08:14:18.9132236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9132300Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9132532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9132636Z layer_outputs = layer_module( 2025-09-07T08:14:18.9132852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9132928Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9133158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9133247Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9133477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.9133566Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9133797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9133865Z return self.weight * hidden_states 2025-09-07T08:14:18.9133869Z 2025-09-07T08:14:18.9134002Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9134193Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9134251Z return mod(**inputs) 2025-09-07T08:14:18.9134522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9134585Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9134819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9134885Z layer_outputs = layer_module( 2025-09-07T08:14:18.9135103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9135171Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9135398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9135488Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9135716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9135830Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9136057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.9136129Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.9136132Z 2025-09-07T08:14:18.9136227Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9136414Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9136476Z return mod(**inputs) 2025-09-07T08:14:18.9136706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9136773Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9137013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9137075Z layer_outputs = layer_module( 2025-09-07T08:14:18.9137344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9137412Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9137643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9137781Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9138010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9138120Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9138348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.9138462Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.9138465Z 2025-09-07T08:14:18.9138552Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9138739Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9138803Z return mod(**inputs) 2025-09-07T08:14:18.9139034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9139102Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9139333Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9139395Z layer_outputs = layer_module( 2025-09-07T08:14:18.9139612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9139712Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9139949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9140027Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9140293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9140394Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9140624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.9140704Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.9140707Z 2025-09-07T08:14:18.9140781Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9140875Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9141057Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9141118Z return mod(**inputs) 2025-09-07T08:14:18.9141351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1725, in forward 2025-09-07T08:14:18.9141414Z encoder_outputs = self.encoder( 2025-09-07T08:14:18.9141652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1128, in forward 2025-09-07T08:14:18.9141748Z hidden_states = self.final_layer_norm(hidden_states) 2025-09-07T08:14:18.9141982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9142052Z return self.weight * hidden_states 2025-09-07T08:14:18.9142055Z 2025-09-07T08:14:18.9142150Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9142343Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9142401Z return mod(**inputs) 2025-09-07T08:14:18.9142646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9142712Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9142945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9143014Z layer_outputs = layer_module( 2025-09-07T08:14:18.9143226Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9143300Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9143573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9143647Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9143883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9143996Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9144229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9144299Z key_states = self.k(current_states) 2025-09-07T08:14:18.9144303Z 2025-09-07T08:14:18.9144401Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9144586Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9144645Z return mod(**inputs) 2025-09-07T08:14:18.9144883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9144949Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9145184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9145247Z layer_outputs = layer_module( 2025-09-07T08:14:18.9145499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9145574Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9145833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9145910Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9146137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9146211Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9146443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9146562Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9146566Z 2025-09-07T08:14:18.9146662Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9146848Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9146913Z return mod(**inputs) 2025-09-07T08:14:18.9147142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9147207Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9147447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9147509Z layer_outputs = layer_module( 2025-09-07T08:14:18.9147728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9147795Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9148024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9148097Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9148326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9148404Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9148631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9148740Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9148749Z 2025-09-07T08:14:18.9148840Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9149024Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9149121Z return mod(**inputs) 2025-09-07T08:14:18.9149351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9149421Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9149692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9149756Z layer_outputs = layer_module( 2025-09-07T08:14:18.9149979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9150046Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9150282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9150353Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9150582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9150657Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9150884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9151050Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9151054Z 2025-09-07T08:14:18.9151142Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9151358Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9151415Z return mod(**inputs) 2025-09-07T08:14:18.9151646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9151716Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9151944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9152013Z layer_outputs = layer_module( 2025-09-07T08:14:18.9152224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9152291Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9152525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9152596Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9152828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9152900Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9153127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9153280Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9153285Z 2025-09-07T08:14:18.9153375Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9153562Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9153619Z return mod(**inputs) 2025-09-07T08:14:18.9153856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9153919Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9154151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9154215Z layer_outputs = layer_module( 2025-09-07T08:14:18.9154428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9154500Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9154727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9154831Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9155062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9155135Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9155408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9155477Z value_states = self.v(current_states) 2025-09-07T08:14:18.9155480Z 2025-09-07T08:14:18.9155575Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9155757Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9155816Z return mod(**inputs) 2025-09-07T08:14:18.9156048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9156112Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9156347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9156407Z layer_outputs = layer_module( 2025-09-07T08:14:18.9156653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9156728Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9156988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9157063Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9157289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9157362Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9157593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9157694Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9157697Z 2025-09-07T08:14:18.9157793Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9157975Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9158038Z return mod(**inputs) 2025-09-07T08:14:18.9158267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9158331Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9158566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9158632Z layer_outputs = layer_module( 2025-09-07T08:14:18.9158849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9158917Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9159144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9159220Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9159447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9159525Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9159750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9159842Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9159851Z 2025-09-07T08:14:18.9159938Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9160121Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9160220Z return mod(**inputs) 2025-09-07T08:14:18.9160450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9160516Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9160745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9160843Z layer_outputs = layer_module( 2025-09-07T08:14:18.9161056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9161122Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9161353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9161424Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9161652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9161730Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9161957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9162053Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9162058Z 2025-09-07T08:14:18.9162177Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9162368Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9162425Z return mod(**inputs) 2025-09-07T08:14:18.9162686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9162760Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9162992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9163061Z layer_outputs = layer_module( 2025-09-07T08:14:18.9163271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9163339Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9163572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9163642Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9163877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9163951Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9164177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9164272Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9164275Z 2025-09-07T08:14:18.9164363Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9164553Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9164608Z return mod(**inputs) 2025-09-07T08:14:18.9164840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9164905Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9165139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9165207Z layer_outputs = layer_module( 2025-09-07T08:14:18.9165419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9165490Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9165719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9165828Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9166062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9166131Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9166358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9166459Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9166462Z 2025-09-07T08:14:18.9166534Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9166623Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9166803Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9166862Z return mod(**inputs) 2025-09-07T08:14:18.9167091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9167157Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9167386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9167445Z layer_outputs = layer_module( 2025-09-07T08:14:18.9167659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9167757Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9167990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9168104Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9168335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.9168424Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9168651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9168725Z return self.weight * hidden_states 2025-09-07T08:14:18.9168728Z 2025-09-07T08:14:18.9168817Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9169003Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9169060Z return mod(**inputs) 2025-09-07T08:14:18.9169288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9169354Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9169583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9169647Z layer_outputs = layer_module( 2025-09-07T08:14:18.9169857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9169925Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9170154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9170231Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9170458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9170568Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9170791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.9170864Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.9170867Z 2025-09-07T08:14:18.9170954Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9171137Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9171192Z return mod(**inputs) 2025-09-07T08:14:18.9171459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9171521Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9171749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9171863Z layer_outputs = layer_module( 2025-09-07T08:14:18.9172073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9172140Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9172366Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9172440Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9172685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9172787Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9173019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.9173091Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.9173094Z 2025-09-07T08:14:18.9173185Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9173407Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9173466Z return mod(**inputs) 2025-09-07T08:14:18.9173732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9173797Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9174033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9174095Z layer_outputs = layer_module( 2025-09-07T08:14:18.9174308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9174382Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9174611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9174693Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9174921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9175023Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9175256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.9175324Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.9175327Z 2025-09-07T08:14:18.9175422Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9175605Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9175667Z return mod(**inputs) 2025-09-07T08:14:18.9175896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9175959Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9176195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9176255Z layer_outputs = layer_module( 2025-09-07T08:14:18.9176470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9176536Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9176764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9176840Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9177104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.9177261Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9177488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9177601Z return self.weight * hidden_states 2025-09-07T08:14:18.9177609Z 2025-09-07T08:14:18.9177698Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9177883Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9177944Z return mod(**inputs) 2025-09-07T08:14:18.9178176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9178244Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9178483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9178548Z layer_outputs = layer_module( 2025-09-07T08:14:18.9178765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9178833Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9179109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9179183Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9179443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9179524Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9179749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9179822Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9179828Z 2025-09-07T08:14:18.9179917Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9180099Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9180162Z return mod(**inputs) 2025-09-07T08:14:18.9180393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9180465Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9180696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9180761Z layer_outputs = layer_module( 2025-09-07T08:14:18.9180971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9181037Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9181268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9181340Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9181572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9181643Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9181872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9181947Z key_states = self.k(current_states) 2025-09-07T08:14:18.9181951Z 2025-09-07T08:14:18.9182040Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9182228Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9182286Z return mod(**inputs) 2025-09-07T08:14:18.9182514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9182620Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9182852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9182920Z layer_outputs = layer_module( 2025-09-07T08:14:18.9183130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9183237Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9183463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9183536Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9183767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9183839Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9184072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9184190Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9184193Z 2025-09-07T08:14:18.9184279Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9184468Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9184558Z return mod(**inputs) 2025-09-07T08:14:18.9184799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9184897Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9185137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9185198Z layer_outputs = layer_module( 2025-09-07T08:14:18.9185415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9185492Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9185720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9185797Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9186023Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9186096Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9186332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9186445Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9186448Z 2025-09-07T08:14:18.9186543Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9186728Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9186787Z return mod(**inputs) 2025-09-07T08:14:18.9187025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9187089Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9187322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9187386Z layer_outputs = layer_module( 2025-09-07T08:14:18.9187605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9187674Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9187901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9187975Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9188203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9188314Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9188544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9188653Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9188688Z 2025-09-07T08:14:18.9188788Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9188972Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9189032Z return mod(**inputs) 2025-09-07T08:14:18.9189263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9189332Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9189563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9189625Z layer_outputs = layer_module( 2025-09-07T08:14:18.9189842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9189911Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9190143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9190245Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9190477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9190584Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9190813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9190965Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9190969Z 2025-09-07T08:14:18.9191059Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9191244Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9191308Z return mod(**inputs) 2025-09-07T08:14:18.9191539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9191610Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9191838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9191904Z layer_outputs = layer_module( 2025-09-07T08:14:18.9192119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9192185Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9192417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9192490Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9192719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9192789Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9193017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9193089Z value_states = self.v(current_states) 2025-09-07T08:14:18.9193092Z 2025-09-07T08:14:18.9193181Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9193371Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9193428Z return mod(**inputs) 2025-09-07T08:14:18.9193658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9193727Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9194013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9194083Z layer_outputs = layer_module( 2025-09-07T08:14:18.9194293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9194408Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9194636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9194708Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9194944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9195015Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9195245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9195344Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9195348Z 2025-09-07T08:14:18.9195438Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9195628Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9195687Z return mod(**inputs) 2025-09-07T08:14:18.9195958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9196024Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9196291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9196355Z layer_outputs = layer_module( 2025-09-07T08:14:18.9196574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9196651Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9196880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9196956Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9197182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9197255Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9197487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9197582Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9197585Z 2025-09-07T08:14:18.9197680Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9197861Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9197918Z return mod(**inputs) 2025-09-07T08:14:18.9198153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9198217Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9198450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9198511Z layer_outputs = layer_module( 2025-09-07T08:14:18.9198734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9198802Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9199031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9199109Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9199334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9199408Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9199671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9199766Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9199769Z 2025-09-07T08:14:18.9199864Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9200083Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9200147Z return mod(**inputs) 2025-09-07T08:14:18.9200380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9200443Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9200679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9200741Z layer_outputs = layer_module( 2025-09-07T08:14:18.9200963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9201032Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9201264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9201333Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9201595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9201674Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9201934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9202034Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9202037Z 2025-09-07T08:14:18.9202125Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9202315Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9202377Z return mod(**inputs) 2025-09-07T08:14:18.9202609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9202679Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9202908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9202980Z layer_outputs = layer_module( 2025-09-07T08:14:18.9203358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9203433Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9203688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9203759Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9203996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9204071Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9204300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9204374Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9204381Z 2025-09-07T08:14:18.9204455Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9204556Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9204755Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9204816Z return mod(**inputs) 2025-09-07T08:14:18.9205057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9205122Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9205361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9205482Z layer_outputs = layer_module( 2025-09-07T08:14:18.9205699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9205778Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9206056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9206135Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9206363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T08:14:18.9206470Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9206698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9206768Z return self.weight * hidden_states 2025-09-07T08:14:18.9206774Z 2025-09-07T08:14:18.9206873Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9207063Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9207127Z return mod(**inputs) 2025-09-07T08:14:18.9207401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9207469Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9207748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9207814Z layer_outputs = layer_module( 2025-09-07T08:14:18.9208036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9208104Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9208337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9208411Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9208641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9208721Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9208955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9209032Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9209035Z 2025-09-07T08:14:18.9209127Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9209314Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9209377Z return mod(**inputs) 2025-09-07T08:14:18.9209611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9209683Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9209913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9209976Z layer_outputs = layer_module( 2025-09-07T08:14:18.9210197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9210268Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9210501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9210574Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9210810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9210884Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9211111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9211220Z key_states = self.k(current_states) 2025-09-07T08:14:18.9211223Z 2025-09-07T08:14:18.9211315Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9211503Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9211594Z return mod(**inputs) 2025-09-07T08:14:18.9211826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9211896Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9212126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9212193Z layer_outputs = layer_module( 2025-09-07T08:14:18.9212406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9212476Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9212710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9212782Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9213050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9213125Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9213360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9213527Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9213530Z 2025-09-07T08:14:18.9213621Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9213813Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9213872Z return mod(**inputs) 2025-09-07T08:14:18.9214110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9214174Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9214404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9214473Z layer_outputs = layer_module( 2025-09-07T08:14:18.9214689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9214764Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9214995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9215065Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9215305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9215380Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9215613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9215725Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9215729Z 2025-09-07T08:14:18.9215828Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9216013Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9216072Z return mod(**inputs) 2025-09-07T08:14:18.9216310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9216373Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9216613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9216707Z layer_outputs = layer_module( 2025-09-07T08:14:18.9216920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9216994Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9217270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9217382Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9217609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9217689Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9217915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9218024Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9218028Z 2025-09-07T08:14:18.9218118Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9218304Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9218363Z return mod(**inputs) 2025-09-07T08:14:18.9218592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9218690Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9218931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9218995Z layer_outputs = layer_module( 2025-09-07T08:14:18.9219245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9219312Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9219542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9219617Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9219843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9219918Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9220144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9220294Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9220298Z 2025-09-07T08:14:18.9220386Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9220568Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9220627Z return mod(**inputs) 2025-09-07T08:14:18.9220857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9220925Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9221155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9221217Z layer_outputs = layer_module( 2025-09-07T08:14:18.9221434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9221504Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9221737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9221810Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9222041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9222114Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9222340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9222455Z value_states = self.v(current_states) 2025-09-07T08:14:18.9222458Z 2025-09-07T08:14:18.9222550Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9222742Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9222835Z return mod(**inputs) 2025-09-07T08:14:18.9223066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9223136Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9223370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9223439Z layer_outputs = layer_module( 2025-09-07T08:14:18.9223652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9223719Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9223954Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9224025Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9224258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9224363Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9224599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9224756Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9224759Z 2025-09-07T08:14:18.9224852Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9225041Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9225097Z return mod(**inputs) 2025-09-07T08:14:18.9225335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9225398Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9225627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9225698Z layer_outputs = layer_module( 2025-09-07T08:14:18.9225911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9225987Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9226220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9226293Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9226526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9226598Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9226835Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9226927Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9226930Z 2025-09-07T08:14:18.9227025Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9227212Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9227270Z return mod(**inputs) 2025-09-07T08:14:18.9227508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9227569Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9227807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9227870Z layer_outputs = layer_module( 2025-09-07T08:14:18.9228317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9228392Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9228620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9228699Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9228966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9229047Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9229275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9229366Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9229370Z 2025-09-07T08:14:18.9229466Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9229650Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9229714Z return mod(**inputs) 2025-09-07T08:14:18.9229943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9230007Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9230278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9230341Z layer_outputs = layer_module( 2025-09-07T08:14:18.9230595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9230665Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9230894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9230974Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9231200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9231279Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9231506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9231609Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9231612Z 2025-09-07T08:14:18.9231701Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9231884Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9231949Z return mod(**inputs) 2025-09-07T08:14:18.9232178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9232249Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9232489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9232556Z layer_outputs = layer_module( 2025-09-07T08:14:18.9232775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9232846Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9233083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9233153Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9233383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9237341Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9237639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9237721Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9237804Z 2025-09-07T08:14:18.9237886Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9237988Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9238202Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9238263Z return mod(**inputs) 2025-09-07T08:14:18.9238564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9238632Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9238879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9238949Z layer_outputs = layer_module( 2025-09-07T08:14:18.9239185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9239264Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9239513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9239602Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9239843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.9239989Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9240243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9240316Z return self.weight * hidden_states 2025-09-07T08:14:18.9240350Z 2025-09-07T08:14:18.9240455Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9240659Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9240723Z return mod(**inputs) 2025-09-07T08:14:18.9240969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9241038Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9241272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9241338Z layer_outputs = layer_module( 2025-09-07T08:14:18.9241559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9241636Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9241867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9241954Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9242224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9242348Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9242590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.9242664Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.9242667Z 2025-09-07T08:14:18.9242767Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9242969Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9243031Z return mod(**inputs) 2025-09-07T08:14:18.9243268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9243335Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9243570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9243634Z layer_outputs = layer_module( 2025-09-07T08:14:18.9243855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9243964Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9244199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9244279Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9244545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9244656Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9244883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.9244959Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.9244962Z 2025-09-07T08:14:18.9245056Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9245246Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9245314Z return mod(**inputs) 2025-09-07T08:14:18.9245545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9245610Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9245870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9245937Z layer_outputs = layer_module( 2025-09-07T08:14:18.9246155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9246255Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9246491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9246569Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9246799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9246903Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9247128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.9247203Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.9247209Z 2025-09-07T08:14:18.9247284Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9247382Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9247575Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9247632Z return mod(**inputs) 2025-09-07T08:14:18.9247867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9247930Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9248166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9248229Z layer_outputs = layer_module( 2025-09-07T08:14:18.9248443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9248516Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9248745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9248821Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9249048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.9249146Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9249373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9249483Z return self.weight * hidden_states 2025-09-07T08:14:18.9249486Z 2025-09-07T08:14:18.9249580Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9249765Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9249828Z return mod(**inputs) 2025-09-07T08:14:18.9250064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9250160Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9250400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9250464Z layer_outputs = layer_module( 2025-09-07T08:14:18.9250680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9250747Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9250975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9251052Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9251294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9251378Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9251642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9251717Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9251721Z 2025-09-07T08:14:18.9251842Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9252032Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9252095Z return mod(**inputs) 2025-09-07T08:14:18.9252326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9252398Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9252627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9252693Z layer_outputs = layer_module( 2025-09-07T08:14:18.9252912Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9252984Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9253223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9253297Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9253529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9253602Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9253828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9253904Z key_states = self.k(current_states) 2025-09-07T08:14:18.9253908Z 2025-09-07T08:14:18.9254002Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9254190Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9254250Z return mod(**inputs) 2025-09-07T08:14:18.9254480Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9254550Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9254781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9254849Z layer_outputs = layer_module( 2025-09-07T08:14:18.9255062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9255166Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9255398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9255470Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9255707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9255810Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9256044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9256169Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9256173Z 2025-09-07T08:14:18.9256264Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9256454Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9256514Z return mod(**inputs) 2025-09-07T08:14:18.9256744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9256807Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9257046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9257150Z layer_outputs = layer_module( 2025-09-07T08:14:18.9257438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9257546Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9257782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9257856Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9258090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9258166Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9258392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9258510Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9258515Z 2025-09-07T08:14:18.9258618Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9258808Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9258864Z return mod(**inputs) 2025-09-07T08:14:18.9259105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9259170Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9259404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9259468Z layer_outputs = layer_module( 2025-09-07T08:14:18.9259683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9259755Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9259984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9260059Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9260283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9260358Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9260583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9260694Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9260697Z 2025-09-07T08:14:18.9260789Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9261010Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9261068Z return mod(**inputs) 2025-09-07T08:14:18.9261299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9261410Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9261640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9261700Z layer_outputs = layer_module( 2025-09-07T08:14:18.9261916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9261982Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9262208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9262281Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9262511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9262581Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9262807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9262990Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9262993Z 2025-09-07T08:14:18.9263084Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9263299Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9263359Z return mod(**inputs) 2025-09-07T08:14:18.9263589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9263653Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9263883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9263943Z layer_outputs = layer_module( 2025-09-07T08:14:18.9264157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9264226Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9264454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9264523Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9264749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9264821Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9265044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9265116Z value_states = self.v(current_states) 2025-09-07T08:14:18.9265120Z 2025-09-07T08:14:18.9265207Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9265392Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9265446Z return mod(**inputs) 2025-09-07T08:14:18.9265677Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9265744Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9265983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9266050Z layer_outputs = layer_module( 2025-09-07T08:14:18.9266262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9266326Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9266603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9266675Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9266909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9267015Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9267250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9267355Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9267358Z 2025-09-07T08:14:18.9267448Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9267642Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9267699Z return mod(**inputs) 2025-09-07T08:14:18.9267937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9268003Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9268232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9268297Z layer_outputs = layer_module( 2025-09-07T08:14:18.9268546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9268624Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9268884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9268955Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9269185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9269257Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9269487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9269579Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9269582Z 2025-09-07T08:14:18.9269674Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9269861Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9269918Z return mod(**inputs) 2025-09-07T08:14:18.9270153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9270214Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9270448Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9270509Z layer_outputs = layer_module( 2025-09-07T08:14:18.9270720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9270789Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9271016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9271089Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9271318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9271387Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9271616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9271703Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9271707Z 2025-09-07T08:14:18.9271799Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9271979Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9272068Z return mod(**inputs) 2025-09-07T08:14:18.9272299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9272360Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9272597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9272694Z layer_outputs = layer_module( 2025-09-07T08:14:18.9272910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9272976Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9273204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9273275Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9273498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9273575Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9273802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9273899Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9273936Z 2025-09-07T08:14:18.9274027Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9274208Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9274302Z return mod(**inputs) 2025-09-07T08:14:18.9274535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9274601Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9274831Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9274896Z layer_outputs = layer_module( 2025-09-07T08:14:18.9275110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9275176Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9275408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9275483Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9275712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9275790Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9276015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9276089Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9276092Z 2025-09-07T08:14:18.9276185Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9276371Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9276428Z return mod(**inputs) 2025-09-07T08:14:18.9276658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9276731Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9276958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9277028Z layer_outputs = layer_module( 2025-09-07T08:14:18.9277239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9277306Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9277533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9277637Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9277872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:14:18.9277997Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:14:18.9278033Z 2025-09-07T08:14:18.9278110Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9278201Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9278387Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9278450Z return mod(**inputs) 2025-09-07T08:14:18.9278682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9278749Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9278981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9279046Z layer_outputs = layer_module( 2025-09-07T08:14:18.9279265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9279332Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9279600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9279676Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9279934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T08:14:18.9280037Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9280265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9280340Z return self.weight * hidden_states 2025-09-07T08:14:18.9280345Z 2025-09-07T08:14:18.9280435Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9280623Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9280680Z return mod(**inputs) 2025-09-07T08:14:18.9280912Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9280984Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9281214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9281283Z layer_outputs = layer_module( 2025-09-07T08:14:18.9281496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9281564Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9281796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9281868Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9282102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9282175Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9282402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9282480Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9282483Z 2025-09-07T08:14:18.9282571Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9282761Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9282817Z return mod(**inputs) 2025-09-07T08:14:18.9283055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9283169Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9283399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9283466Z layer_outputs = layer_module( 2025-09-07T08:14:18.9283678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9283785Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9284011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9284083Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9284316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9284390Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9284622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9284693Z key_states = self.k(current_states) 2025-09-07T08:14:18.9284697Z 2025-09-07T08:14:18.9284785Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9284969Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9285027Z return mod(**inputs) 2025-09-07T08:14:18.9285293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9285357Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9285618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9285679Z layer_outputs = layer_module( 2025-09-07T08:14:18.9285890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9285963Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9286191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9286265Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9286495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9286568Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9286802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9286923Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9286927Z 2025-09-07T08:14:18.9287019Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9287200Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9287256Z return mod(**inputs) 2025-09-07T08:14:18.9287490Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9287553Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9287784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9287847Z layer_outputs = layer_module( 2025-09-07T08:14:18.9288068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9288135Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9288368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9288439Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9288664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9288772Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9288997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9289111Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9289116Z 2025-09-07T08:14:18.9289213Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9289435Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9289497Z return mod(**inputs) 2025-09-07T08:14:18.9289733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9289797Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9290033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9290096Z layer_outputs = layer_module( 2025-09-07T08:14:18.9290317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9290382Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9290608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9290713Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9290947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9291021Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9291276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9291394Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9291397Z 2025-09-07T08:14:18.9291486Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9291680Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9291738Z return mod(**inputs) 2025-09-07T08:14:18.9291969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9292037Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9292272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9292338Z layer_outputs = layer_module( 2025-09-07T08:14:18.9292552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9292620Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9292852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9292923Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9293155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9293227Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9293459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9293610Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9293613Z 2025-09-07T08:14:18.9293704Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9293894Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9293953Z return mod(**inputs) 2025-09-07T08:14:18.9294186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9294250Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9294519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9294586Z layer_outputs = layer_module( 2025-09-07T08:14:18.9294799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9294870Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9295131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9295204Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9295437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9295510Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9295743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9295814Z value_states = self.v(current_states) 2025-09-07T08:14:18.9295818Z 2025-09-07T08:14:18.9295910Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9296092Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9296150Z return mod(**inputs) 2025-09-07T08:14:18.9296414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9296479Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9296751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9296814Z layer_outputs = layer_module( 2025-09-07T08:14:18.9297026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9297099Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9297392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9297468Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9297693Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9297773Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9298002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9298100Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9298103Z 2025-09-07T08:14:18.9298204Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9298389Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9298451Z return mod(**inputs) 2025-09-07T08:14:18.9298681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9298747Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9298984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9299045Z layer_outputs = layer_module( 2025-09-07T08:14:18.9299260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9299329Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9299560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9299637Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9299862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9299941Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9300211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9300308Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9300311Z 2025-09-07T08:14:18.9300398Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9300585Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9300683Z return mod(**inputs) 2025-09-07T08:14:18.9300915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9300979Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9301209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9301270Z layer_outputs = layer_module( 2025-09-07T08:14:18.9301489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9301556Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9301788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9301859Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9302122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9302201Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9302612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9302706Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9302710Z 2025-09-07T08:14:18.9302799Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9302986Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9303273Z return mod(**inputs) 2025-09-07T08:14:18.9303535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9303602Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9303837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9303905Z layer_outputs = layer_module( 2025-09-07T08:14:18.9304117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9304182Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9304413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9304481Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9304714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9304785Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9305017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9305117Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9305124Z 2025-09-07T08:14:18.9305215Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9305399Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9305456Z return mod(**inputs) 2025-09-07T08:14:18.9305693Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9305756Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9305987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9306125Z layer_outputs = layer_module( 2025-09-07T08:14:18.9306338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9306406Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9306637Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9306755Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9306989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9307062Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9307292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9307361Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9307364Z 2025-09-07T08:14:18.9307439Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9307528Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9307710Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9307773Z return mod(**inputs) 2025-09-07T08:14:18.9308047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9308119Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9308348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9308455Z layer_outputs = layer_module( 2025-09-07T08:14:18.9308672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9308741Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9308975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9309058Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9309285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.9309376Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9309606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9309679Z return self.weight * hidden_states 2025-09-07T08:14:18.9309682Z 2025-09-07T08:14:18.9309775Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9309960Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9310016Z return mod(**inputs) 2025-09-07T08:14:18.9310245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9310316Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9310545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9310612Z layer_outputs = layer_module( 2025-09-07T08:14:18.9310823Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9310892Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9311124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9311205Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9311437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9311546Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9311786Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.9311899Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.9311902Z 2025-09-07T08:14:18.9311993Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9312181Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9312271Z return mod(**inputs) 2025-09-07T08:14:18.9312506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9312572Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9312803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9312870Z layer_outputs = layer_module( 2025-09-07T08:14:18.9313080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9313152Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9313381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9313459Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9313721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9313825Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9314085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.9314160Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.9314163Z 2025-09-07T08:14:18.9314254Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9314438Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9314495Z return mod(**inputs) 2025-09-07T08:14:18.9314730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9314790Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9315024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9315088Z layer_outputs = layer_module( 2025-09-07T08:14:18.9315300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9315374Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9315602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9315682Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9315905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9316006Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9316237Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.9316304Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.9316309Z 2025-09-07T08:14:18.9316386Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9316475Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9316664Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9316722Z return mod(**inputs) 2025-09-07T08:14:18.9316952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9317018Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9317247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9317348Z layer_outputs = layer_module( 2025-09-07T08:14:18.9317561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9317628Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9317871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9317976Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9318207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.9318305Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9318531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9318605Z return self.weight * hidden_states 2025-09-07T08:14:18.9318609Z 2025-09-07T08:14:18.9318699Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9318887Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9318944Z return mod(**inputs) 2025-09-07T08:14:18.9319175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9319311Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9319546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9319641Z layer_outputs = layer_module( 2025-09-07T08:14:18.9319866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9319940Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9320167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9320241Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9320479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9320555Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9320799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9320868Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9320872Z 2025-09-07T08:14:18.9320965Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9321148Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9321203Z return mod(**inputs) 2025-09-07T08:14:18.9321435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9321498Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9321731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9321793Z layer_outputs = layer_module( 2025-09-07T08:14:18.9322003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9322076Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9322304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9322377Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9322600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9322669Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9322898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9323000Z key_states = self.k(current_states) 2025-09-07T08:14:18.9323003Z 2025-09-07T08:14:18.9323096Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9323278Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9323368Z return mod(**inputs) 2025-09-07T08:14:18.9323601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9323664Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9323900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9323959Z layer_outputs = layer_module( 2025-09-07T08:14:18.9324173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9324240Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9324468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9324542Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9324769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9324893Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9325122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9325269Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9325277Z 2025-09-07T08:14:18.9325364Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9325544Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9325604Z return mod(**inputs) 2025-09-07T08:14:18.9325834Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9325902Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9326129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9326192Z layer_outputs = layer_module( 2025-09-07T08:14:18.9326408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9326475Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9326707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9326774Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9327000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9327076Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9327304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9327419Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9327422Z 2025-09-07T08:14:18.9327509Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9327696Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9327752Z return mod(**inputs) 2025-09-07T08:14:18.9327981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9328047Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9328276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9328342Z layer_outputs = layer_module( 2025-09-07T08:14:18.9328593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9328661Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9328893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9328998Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9329229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9329298Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9329524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9329640Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9329643Z 2025-09-07T08:14:18.9329731Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9329919Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9329973Z return mod(**inputs) 2025-09-07T08:14:18.9330203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9330263Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9330529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9330597Z layer_outputs = layer_module( 2025-09-07T08:14:18.9330840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9330911Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9331139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9331208Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9331443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9331512Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9331742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9331891Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9331895Z 2025-09-07T08:14:18.9331988Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9332173Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9332230Z return mod(**inputs) 2025-09-07T08:14:18.9332465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9332526Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9332762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9332824Z layer_outputs = layer_module( 2025-09-07T08:14:18.9333035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9333102Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9333330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9333406Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9333644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9333715Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9333942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9334011Z value_states = self.v(current_states) 2025-09-07T08:14:18.9334049Z 2025-09-07T08:14:18.9334145Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9334327Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9334386Z return mod(**inputs) 2025-09-07T08:14:18.9334617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9334711Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9334951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9335014Z layer_outputs = layer_module( 2025-09-07T08:14:18.9335233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9335299Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9335523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9335600Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9335829Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9335904Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9336163Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9336263Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9336271Z 2025-09-07T08:14:18.9336389Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9336573Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9336636Z return mod(**inputs) 2025-09-07T08:14:18.9336867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9336932Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9337221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9337284Z layer_outputs = layer_module( 2025-09-07T08:14:18.9337498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9337565Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9337799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9337868Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9338094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9338170Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9338401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9338502Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9338504Z 2025-09-07T08:14:18.9338598Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9338790Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9338855Z return mod(**inputs) 2025-09-07T08:14:18.9339092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9339161Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9339394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9339463Z layer_outputs = layer_module( 2025-09-07T08:14:18.9339679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9339790Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9340027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9340099Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9340335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9340440Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9340669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9340770Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9340774Z 2025-09-07T08:14:18.9340863Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9341052Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9341111Z return mod(**inputs) 2025-09-07T08:14:18.9341346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9341411Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9341641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9341746Z layer_outputs = layer_module( 2025-09-07T08:14:18.9341963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9342063Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9342295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9342366Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9342600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9342672Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9342903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9342996Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9343001Z 2025-09-07T08:14:18.9343093Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9343287Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9343343Z return mod(**inputs) 2025-09-07T08:14:18.9343582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9343651Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9343880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9343941Z layer_outputs = layer_module( 2025-09-07T08:14:18.9344151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9344219Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9344445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9344517Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9344741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9344809Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9345037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9345104Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9345107Z 2025-09-07T08:14:18.9345181Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9345303Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9345484Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9345542Z return mod(**inputs) 2025-09-07T08:14:18.9345770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9345887Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9346116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9346179Z layer_outputs = layer_module( 2025-09-07T08:14:18.9346389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9346455Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9346685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9346755Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9346982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T08:14:18.9347076Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9347340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9347410Z return self.weight * hidden_states 2025-09-07T08:14:18.9347414Z 2025-09-07T08:14:18.9347499Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9347723Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9347779Z return mod(**inputs) 2025-09-07T08:14:18.9348009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9348077Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9348307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9348369Z layer_outputs = layer_module( 2025-09-07T08:14:18.9348581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9348650Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9348877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9348950Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9349178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9349251Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9349481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9349549Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9349552Z 2025-09-07T08:14:18.9349640Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9349825Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9349883Z return mod(**inputs) 2025-09-07T08:14:18.9350114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9350176Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9350428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9350492Z layer_outputs = layer_module( 2025-09-07T08:14:18.9350704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9350776Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9351040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9351112Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9351336Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9351445Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9351674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9351743Z key_states = self.k(current_states) 2025-09-07T08:14:18.9351747Z 2025-09-07T08:14:18.9351839Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9352020Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9352074Z return mod(**inputs) 2025-09-07T08:14:18.9352304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9352367Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9352598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9352657Z layer_outputs = layer_module( 2025-09-07T08:14:18.9352903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9352971Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9353242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9353314Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9353540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9353612Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9353839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9353956Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9353960Z 2025-09-07T08:14:18.9354051Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9354235Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9354292Z return mod(**inputs) 2025-09-07T08:14:18.9354521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9354582Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9354813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9354873Z layer_outputs = layer_module( 2025-09-07T08:14:18.9355085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9355150Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9355379Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9355447Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9355677Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9355749Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9355973Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9356083Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9356087Z 2025-09-07T08:14:18.9356173Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9356353Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9356446Z return mod(**inputs) 2025-09-07T08:14:18.9356675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9356738Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9357005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9357066Z layer_outputs = layer_module( 2025-09-07T08:14:18.9357281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9357346Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9357573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9357641Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9357869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9357940Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9358164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9358308Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9358311Z 2025-09-07T08:14:18.9358399Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9358621Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9358682Z return mod(**inputs) 2025-09-07T08:14:18.9358922Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9358991Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9359222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9359290Z layer_outputs = layer_module( 2025-09-07T08:14:18.9359506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9359580Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9359817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9359889Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9360119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9360191Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9360418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9360566Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9360571Z 2025-09-07T08:14:18.9360679Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9360875Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9360932Z return mod(**inputs) 2025-09-07T08:14:18.9361168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9361231Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9361461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9361525Z layer_outputs = layer_module( 2025-09-07T08:14:18.9361739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9361807Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9362031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9362137Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9362366Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9362436Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9362698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9362769Z value_states = self.v(current_states) 2025-09-07T08:14:18.9362773Z 2025-09-07T08:14:18.9362868Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9363058Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9363118Z return mod(**inputs) 2025-09-07T08:14:18.9363361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9363430Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9363668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9363731Z layer_outputs = layer_module( 2025-09-07T08:14:18.9363981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9364058Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9364323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9364397Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9364625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9364700Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9364924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9365027Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9365030Z 2025-09-07T08:14:18.9365129Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9365323Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9365383Z return mod(**inputs) 2025-09-07T08:14:18.9365615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9365679Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9365908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9365971Z layer_outputs = layer_module( 2025-09-07T08:14:18.9366187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9366255Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9366483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9366552Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9366777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9366854Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9367085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9367179Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9367182Z 2025-09-07T08:14:18.9367272Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9367465Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9367578Z return mod(**inputs) 2025-09-07T08:14:18.9367810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9367875Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9368110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9368209Z layer_outputs = layer_module( 2025-09-07T08:14:18.9368430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9368499Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9368734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9368804Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9369036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9369111Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9369339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9369443Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9369448Z 2025-09-07T08:14:18.9369578Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9369769Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9369826Z return mod(**inputs) 2025-09-07T08:14:18.9370090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9370163Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9370396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9370467Z layer_outputs = layer_module( 2025-09-07T08:14:18.9370681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9370752Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9370988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9371062Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9371295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9371367Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9371604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9371701Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9371704Z 2025-09-07T08:14:18.9371797Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9371993Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9372053Z return mod(**inputs) 2025-09-07T08:14:18.9372293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9372360Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9372610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9372680Z layer_outputs = layer_module( 2025-09-07T08:14:18.9372895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9372971Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9373197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9373307Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9373536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9373610Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9373844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9373952Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9373956Z 2025-09-07T08:14:18.9374052Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9374239Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9374295Z return mod(**inputs) 2025-09-07T08:14:18.9374531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9374595Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9374830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9374892Z layer_outputs = layer_module( 2025-09-07T08:14:18.9375106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9375219Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9375451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9375529Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9375789Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 647, in forward 2025-09-07T08:14:18.9375921Z layer_output = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:14:18.9375924Z 2025-09-07T08:14:18.9375998Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9376094Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9376287Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9376346Z return mod(**inputs) 2025-09-07T08:14:18.9376583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9376650Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9376880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9376949Z layer_outputs = layer_module( 2025-09-07T08:14:18.9377203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9377282Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9377513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9377598Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9377832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.9377922Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9378154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9378225Z return self.weight * hidden_states 2025-09-07T08:14:18.9378228Z 2025-09-07T08:14:18.9378321Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9378507Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9378565Z return mod(**inputs) 2025-09-07T08:14:18.9378802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9378866Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9379146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9379211Z layer_outputs = layer_module( 2025-09-07T08:14:18.9379424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9379533Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9379762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9379849Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9380075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9380186Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9380419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.9380495Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.9380498Z 2025-09-07T08:14:18.9380592Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9380776Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9380841Z return mod(**inputs) 2025-09-07T08:14:18.9381105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9381171Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9381443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9381507Z layer_outputs = layer_module( 2025-09-07T08:14:18.9381723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9381792Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9382022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9382104Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9382331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9382447Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9382673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.9382753Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.9382757Z 2025-09-07T08:14:18.9382847Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9383032Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9383095Z return mod(**inputs) 2025-09-07T08:14:18.9383325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9383396Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9383626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9383689Z layer_outputs = layer_module( 2025-09-07T08:14:18.9383910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9383979Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9384209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9384282Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9384510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9384652Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9384879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.9384954Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.9384957Z 2025-09-07T08:14:18.9385027Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9385160Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9385343Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9385401Z return mod(**inputs) 2025-09-07T08:14:18.9385644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9385709Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9385946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9386011Z layer_outputs = layer_module( 2025-09-07T08:14:18.9386224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9386297Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9386527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9386641Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9386876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.9387021Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9387256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9387325Z return self.weight * hidden_states 2025-09-07T08:14:18.9387328Z 2025-09-07T08:14:18.9387426Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9387609Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9387674Z return mod(**inputs) 2025-09-07T08:14:18.9387903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9387970Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9388208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9388270Z layer_outputs = layer_module( 2025-09-07T08:14:18.9388488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9388555Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9388782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9388861Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9389090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9389170Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9389399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9389471Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9389480Z 2025-09-07T08:14:18.9389568Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9389753Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9389819Z return mod(**inputs) 2025-09-07T08:14:18.9390050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9390118Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9390386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9390449Z layer_outputs = layer_module( 2025-09-07T08:14:18.9390669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9390737Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9391008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9391077Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9391304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9391375Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9391601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9391673Z key_states = self.k(current_states) 2025-09-07T08:14:18.9391676Z 2025-09-07T08:14:18.9391763Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9391945Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9392000Z return mod(**inputs) 2025-09-07T08:14:18.9392259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9392324Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9392586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9392649Z layer_outputs = layer_module( 2025-09-07T08:14:18.9392859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9392925Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9393151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9393222Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9393453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9393520Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9393748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9393870Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9393874Z 2025-09-07T08:14:18.9393964Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9394147Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9394202Z return mod(**inputs) 2025-09-07T08:14:18.9394431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9394493Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9394720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9394782Z layer_outputs = layer_module( 2025-09-07T08:14:18.9394994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9395061Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9395288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9395357Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9395585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9395654Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9395917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9396030Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9396034Z 2025-09-07T08:14:18.9396126Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9396312Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9396398Z return mod(**inputs) 2025-09-07T08:14:18.9396631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9396701Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9396939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9397003Z layer_outputs = layer_module( 2025-09-07T08:14:18.9397220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9397294Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9397524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9397601Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9397862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9397938Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9398212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9398324Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9398327Z 2025-09-07T08:14:18.9398425Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9398607Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9398674Z return mod(**inputs) 2025-09-07T08:14:18.9398905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9398969Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9399209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9399271Z layer_outputs = layer_module( 2025-09-07T08:14:18.9399485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9399554Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9399784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9399863Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9400091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9400171Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9400404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9400551Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9400564Z 2025-09-07T08:14:18.9400656Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9400841Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9400907Z return mod(**inputs) 2025-09-07T08:14:18.9401136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9401207Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9401437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9401535Z layer_outputs = layer_module( 2025-09-07T08:14:18.9401755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9401822Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9402057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9402165Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9402396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9402477Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9402704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9402778Z value_states = self.v(current_states) 2025-09-07T08:14:18.9402783Z 2025-09-07T08:14:18.9402875Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9403205Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9403262Z return mod(**inputs) 2025-09-07T08:14:18.9403494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9403620Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9403864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9403976Z layer_outputs = layer_module( 2025-09-07T08:14:18.9404203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9404279Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9404519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9404596Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9404824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9404895Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9405121Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9405226Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9405229Z 2025-09-07T08:14:18.9405327Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9405518Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9405574Z return mod(**inputs) 2025-09-07T08:14:18.9405814Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9405883Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9406112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9406179Z layer_outputs = layer_module( 2025-09-07T08:14:18.9406396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9406470Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9406701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9406775Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9407008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9407077Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9407313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9407454Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9407458Z 2025-09-07T08:14:18.9407557Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9407741Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9407869Z return mod(**inputs) 2025-09-07T08:14:18.9408109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9408169Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9408405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9408465Z layer_outputs = layer_module( 2025-09-07T08:14:18.9408683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9408767Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9408999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9409074Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9409300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9409422Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9409659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9409782Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9409786Z 2025-09-07T08:14:18.9409885Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9410069Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9410135Z return mod(**inputs) 2025-09-07T08:14:18.9410371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9410432Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9410667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9410729Z layer_outputs = layer_module( 2025-09-07T08:14:18.9410951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9411016Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9411247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9411318Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9411544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9411618Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9411847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9411939Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9411947Z 2025-09-07T08:14:18.9412035Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9412220Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9412285Z return mod(**inputs) 2025-09-07T08:14:18.9412520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9412590Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9412820Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9412883Z layer_outputs = layer_module( 2025-09-07T08:14:18.9413142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9413209Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9413439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9413541Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9413768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9413842Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9414069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9414143Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9414146Z 2025-09-07T08:14:18.9414218Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9414307Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9414496Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9414553Z return mod(**inputs) 2025-09-07T08:14:18.9414787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9414851Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9415114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9415177Z layer_outputs = layer_module( 2025-09-07T08:14:18.9415420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9415492Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9415718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9415796Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9416022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T08:14:18.9416119Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9416359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9416430Z return self.weight * hidden_states 2025-09-07T08:14:18.9416433Z 2025-09-07T08:14:18.9416532Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9416720Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9416780Z return mod(**inputs) 2025-09-07T08:14:18.9417020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9417084Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9417383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9417448Z layer_outputs = layer_module( 2025-09-07T08:14:18.9417675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9417748Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9417982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9418055Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9418282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9418356Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9418580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9418690Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9418693Z 2025-09-07T08:14:18.9418784Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9418966Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9419027Z return mod(**inputs) 2025-09-07T08:14:18.9419292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9419357Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9419591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9419656Z layer_outputs = layer_module( 2025-09-07T08:14:18.9419883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9419954Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9420185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9420255Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9420480Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9420560Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9420820Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9420894Z key_states = self.k(current_states) 2025-09-07T08:14:18.9420897Z 2025-09-07T08:14:18.9421017Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9421203Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9421269Z return mod(**inputs) 2025-09-07T08:14:18.9421498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9421563Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9421790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9421850Z layer_outputs = layer_module( 2025-09-07T08:14:18.9422063Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9422129Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9422363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9422432Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9422662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9422732Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9422956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9423081Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9423085Z 2025-09-07T08:14:18.9423173Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9423364Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9423425Z return mod(**inputs) 2025-09-07T08:14:18.9423652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9423721Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9423951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9424019Z layer_outputs = layer_module( 2025-09-07T08:14:18.9424229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9424332Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9424558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9424625Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9424892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9424964Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9425196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9425310Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9425313Z 2025-09-07T08:14:18.9425404Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9425595Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9425654Z return mod(**inputs) 2025-09-07T08:14:18.9425894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9425957Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9426218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9426285Z layer_outputs = layer_module( 2025-09-07T08:14:18.9426496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9426598Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9426827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9426899Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9427124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9427196Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9427422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9427530Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9427536Z 2025-09-07T08:14:18.9427630Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9427816Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9427874Z return mod(**inputs) 2025-09-07T08:14:18.9428109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9428170Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9428403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9428464Z layer_outputs = layer_module( 2025-09-07T08:14:18.9428681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9428745Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9428972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9429046Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9429283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9429358Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9429583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9429726Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9429768Z 2025-09-07T08:14:18.9429859Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9430052Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9430116Z return mod(**inputs) 2025-09-07T08:14:18.9430351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9430464Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9430698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9430759Z layer_outputs = layer_module( 2025-09-07T08:14:18.9430974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9431041Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9431278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9431350Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9431577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9431654Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9431916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9431996Z value_states = self.v(current_states) 2025-09-07T08:14:18.9431999Z 2025-09-07T08:14:18.9432118Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9432309Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9432373Z return mod(**inputs) 2025-09-07T08:14:18.9432607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9432679Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9432907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9432967Z layer_outputs = layer_module( 2025-09-07T08:14:18.9433185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9433254Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9433486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9433555Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9433785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9433858Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9434083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9434189Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9434192Z 2025-09-07T08:14:18.9434282Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9434471Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9434532Z return mod(**inputs) 2025-09-07T08:14:18.9434762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9434834Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9435065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9435135Z layer_outputs = layer_module( 2025-09-07T08:14:18.9435348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9435459Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9435688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9435756Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9435988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9436094Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9436326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9436417Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9436421Z 2025-09-07T08:14:18.9436508Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9436691Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9436746Z return mod(**inputs) 2025-09-07T08:14:18.9436978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9437038Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9437264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9437364Z layer_outputs = layer_module( 2025-09-07T08:14:18.9437588Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9437662Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9437927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9438008Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9438239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9438314Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9438542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9438637Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9438640Z 2025-09-07T08:14:18.9438739Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9438930Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9438991Z return mod(**inputs) 2025-09-07T08:14:18.9439236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9439302Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9439536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9439601Z layer_outputs = layer_module( 2025-09-07T08:14:18.9439819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9439887Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9440113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9440187Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9440412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9440489Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9440719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9440815Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9440818Z 2025-09-07T08:14:18.9440915Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9441136Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9441204Z return mod(**inputs) 2025-09-07T08:14:18.9441437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9441502Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9441787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9441855Z layer_outputs = layer_module( 2025-09-07T08:14:18.9442075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9442142Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9442372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9442443Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9442669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9442741Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9442966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9443075Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9443079Z 2025-09-07T08:14:18.9443153Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9443244Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9443465Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9443524Z return mod(**inputs) 2025-09-07T08:14:18.9443762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9443823Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9444054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9444116Z layer_outputs = layer_module( 2025-09-07T08:14:18.9444326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9444397Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9444623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9444709Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9444935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.9445021Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9445251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9445320Z return self.weight * hidden_states 2025-09-07T08:14:18.9445323Z 2025-09-07T08:14:18.9445412Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9445596Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9445653Z return mod(**inputs) 2025-09-07T08:14:18.9445890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9445952Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9446188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9446247Z layer_outputs = layer_module( 2025-09-07T08:14:18.9446457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9446524Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9446787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9446872Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9447098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9447449Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9447680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.9447755Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.9447759Z 2025-09-07T08:14:18.9447855Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9448037Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9448099Z return mod(**inputs) 2025-09-07T08:14:18.9448327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9448393Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9448623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9448684Z layer_outputs = layer_module( 2025-09-07T08:14:18.9448937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9449002Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9449262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9449342Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9449568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9449678Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9449905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.9449983Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.9449987Z 2025-09-07T08:14:18.9450077Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9450265Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9450323Z return mod(**inputs) 2025-09-07T08:14:18.9450562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9450633Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9450862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9450924Z layer_outputs = layer_module( 2025-09-07T08:14:18.9451138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9451207Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9451442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9455818Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9456122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9456243Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9456499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.9456579Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.9456583Z 2025-09-07T08:14:18.9456686Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9456895Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9457024Z return mod(**inputs) 2025-09-07T08:14:18.9457343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9457422Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9457706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9457779Z layer_outputs = layer_module( 2025-09-07T08:14:18.9458009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9458086Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9458328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9458415Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9458651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 343, in forward 2025-09-07T08:14:18.9458773Z hidden_states = hidden_states + self.dropout(forwarded_states) 2025-09-07T08:14:18.9458777Z 2025-09-07T08:14:18.9458854Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9458956Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9459191Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9459254Z return mod(**inputs) 2025-09-07T08:14:18.9459552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9459621Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9459857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9459921Z layer_outputs = layer_module( 2025-09-07T08:14:18.9460150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9460221Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9460454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9460531Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9460761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.9460861Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9461091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9461164Z return self.weight * hidden_states 2025-09-07T08:14:18.9461168Z 2025-09-07T08:14:18.9461266Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9461467Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9461527Z return mod(**inputs) 2025-09-07T08:14:18.9461761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9461829Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9462062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9462128Z layer_outputs = layer_module( 2025-09-07T08:14:18.9462347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9462415Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9462648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9462720Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9462996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9463076Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9463306Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9463415Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9463418Z 2025-09-07T08:14:18.9463509Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9463703Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9463760Z return mod(**inputs) 2025-09-07T08:14:18.9463990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9464057Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9464288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9464357Z layer_outputs = layer_module( 2025-09-07T08:14:18.9464569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9464639Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9464908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9464981Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9465247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9465322Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9465553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9465619Z key_states = self.k(current_states) 2025-09-07T08:14:18.9465624Z 2025-09-07T08:14:18.9465712Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9465899Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9465955Z return mod(**inputs) 2025-09-07T08:14:18.9466189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9466251Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9466482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9466550Z layer_outputs = layer_module( 2025-09-07T08:14:18.9466764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9466837Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9467064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9467133Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9467366Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9467434Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9467663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9467784Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9467788Z 2025-09-07T08:14:18.9467879Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9468063Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9468118Z return mod(**inputs) 2025-09-07T08:14:18.9468353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9468453Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9468690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9468756Z layer_outputs = layer_module( 2025-09-07T08:14:18.9468975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9469080Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9469308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9469389Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9469613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9469688Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9469913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9470028Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9470032Z 2025-09-07T08:14:18.9470130Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9470315Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9470410Z return mod(**inputs) 2025-09-07T08:14:18.9470645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9470741Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9470976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9471038Z layer_outputs = layer_module( 2025-09-07T08:14:18.9471255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9471322Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9471548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9471622Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9471848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9471925Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9472150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9472263Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9472267Z 2025-09-07T08:14:18.9472354Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9472540Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9472610Z return mod(**inputs) 2025-09-07T08:14:18.9472843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9472912Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9473140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9473205Z layer_outputs = layer_module( 2025-09-07T08:14:18.9473421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9473489Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9473719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9473786Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9474016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9474119Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9474346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9474499Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9474536Z 2025-09-07T08:14:18.9474630Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9474825Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9474882Z return mod(**inputs) 2025-09-07T08:14:18.9475130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9475197Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9475427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9475494Z layer_outputs = layer_module( 2025-09-07T08:14:18.9475709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9475778Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9476015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9476141Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9476377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9476479Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9476715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9476784Z value_states = self.v(current_states) 2025-09-07T08:14:18.9476787Z 2025-09-07T08:14:18.9476880Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9477072Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9477130Z return mod(**inputs) 2025-09-07T08:14:18.9477364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9477430Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9477661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9477727Z layer_outputs = layer_module( 2025-09-07T08:14:18.9477948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9478022Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9478254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9478323Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9478553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9478624Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9478853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9478957Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9478961Z 2025-09-07T08:14:18.9479058Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9479246Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9479306Z return mod(**inputs) 2025-09-07T08:14:18.9479539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9479602Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9479881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9479944Z layer_outputs = layer_module( 2025-09-07T08:14:18.9480154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9480257Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9480486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9480557Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9480786Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9480865Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9481095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9481188Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9481192Z 2025-09-07T08:14:18.9481286Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9481468Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9481530Z return mod(**inputs) 2025-09-07T08:14:18.9481794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9481861Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9482126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9482190Z layer_outputs = layer_module( 2025-09-07T08:14:18.9482410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9482478Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9482706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9482778Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9483004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9483085Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9483309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9483401Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9483407Z 2025-09-07T08:14:18.9483498Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9483682Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9483741Z return mod(**inputs) 2025-09-07T08:14:18.9483975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9484044Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9484275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9484334Z layer_outputs = layer_module( 2025-09-07T08:14:18.9484554Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9484620Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9484854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9484924Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9485149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9485222Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9485483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9485579Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9485582Z 2025-09-07T08:14:18.9485671Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9485896Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9485954Z return mod(**inputs) 2025-09-07T08:14:18.9486187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9486250Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9486483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9486546Z layer_outputs = layer_module( 2025-09-07T08:14:18.9486755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9486821Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9487047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9487117Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9487386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9487458Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9487727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9487798Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9487802Z 2025-09-07T08:14:18.9487878Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9487977Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9488161Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9488224Z return mod(**inputs) 2025-09-07T08:14:18.9488455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9488519Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9488757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9488818Z layer_outputs = layer_module( 2025-09-07T08:14:18.9489039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9489108Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9489334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9489411Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9489638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T08:14:18.9489743Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9489968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9490044Z return self.weight * hidden_states 2025-09-07T08:14:18.9490047Z 2025-09-07T08:14:18.9490141Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9490328Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9490387Z return mod(**inputs) 2025-09-07T08:14:18.9490616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9490679Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9490907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9491003Z layer_outputs = layer_module( 2025-09-07T08:14:18.9491220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9491287Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9491552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9491623Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9491849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9491926Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9492150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9492223Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9492228Z 2025-09-07T08:14:18.9492319Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9492504Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9492559Z return mod(**inputs) 2025-09-07T08:14:18.9492819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9492889Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9493117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9493212Z layer_outputs = layer_module( 2025-09-07T08:14:18.9493427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9493492Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9493725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9493796Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9494024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9494097Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9494325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9494399Z key_states = self.k(current_states) 2025-09-07T08:14:18.9494402Z 2025-09-07T08:14:18.9494498Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9494690Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9494747Z return mod(**inputs) 2025-09-07T08:14:18.9494982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9495046Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9495277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9495344Z layer_outputs = layer_module( 2025-09-07T08:14:18.9495555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9495631Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9495859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9495931Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9496164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9496238Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9496473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9496630Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9496633Z 2025-09-07T08:14:18.9496727Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9496911Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9497018Z return mod(**inputs) 2025-09-07T08:14:18.9497293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9497356Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9497593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9497655Z layer_outputs = layer_module( 2025-09-07T08:14:18.9497868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9497939Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9498166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9498243Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9498509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9498584Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9498812Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9498956Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9498960Z 2025-09-07T08:14:18.9499054Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9499236Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9499298Z return mod(**inputs) 2025-09-07T08:14:18.9499527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9499588Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9499821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9499885Z layer_outputs = layer_module( 2025-09-07T08:14:18.9500103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9500171Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9500395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9500467Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9500693Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9500773Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9500996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9501103Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9501112Z 2025-09-07T08:14:18.9501204Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9501385Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9501447Z return mod(**inputs) 2025-09-07T08:14:18.9501676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9501743Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9501974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9502097Z layer_outputs = layer_module( 2025-09-07T08:14:18.9502311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9502377Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9502611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9502715Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9502941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9503020Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9503411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9503565Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9503569Z 2025-09-07T08:14:18.9503659Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9503850Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9503906Z return mod(**inputs) 2025-09-07T08:14:18.9504139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9504268Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9504508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9504574Z layer_outputs = layer_module( 2025-09-07T08:14:18.9504837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9504911Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9505146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9505218Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9505451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9505521Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9505747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9505819Z value_states = self.v(current_states) 2025-09-07T08:14:18.9505823Z 2025-09-07T08:14:18.9505917Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9506108Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9506164Z return mod(**inputs) 2025-09-07T08:14:18.9506398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9506463Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9506692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9506757Z layer_outputs = layer_module( 2025-09-07T08:14:18.9506971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9507041Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9507267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9507334Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9507564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9507635Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9507860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9508006Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9508009Z 2025-09-07T08:14:18.9508104Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9508292Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9508348Z return mod(**inputs) 2025-09-07T08:14:18.9508634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9508702Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9508943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9509005Z layer_outputs = layer_module( 2025-09-07T08:14:18.9509218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9509292Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9509523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9509603Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9509833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9509946Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9510175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9510305Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9510309Z 2025-09-07T08:14:18.9510406Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9510594Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9510655Z return mod(**inputs) 2025-09-07T08:14:18.9510888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9510952Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9511190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9511252Z layer_outputs = layer_module( 2025-09-07T08:14:18.9511475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9511543Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9511776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9511846Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9512072Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9512145Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9512374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9512470Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9512473Z 2025-09-07T08:14:18.9512560Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9512762Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9512823Z return mod(**inputs) 2025-09-07T08:14:18.9513054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9513122Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9513350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9513411Z layer_outputs = layer_module( 2025-09-07T08:14:18.9513624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9513736Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9513968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9514036Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9514305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9514378Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9514606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9514701Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9514705Z 2025-09-07T08:14:18.9514792Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9514977Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9515033Z return mod(**inputs) 2025-09-07T08:14:18.9515264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9515327Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9515594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9515662Z layer_outputs = layer_module( 2025-09-07T08:14:18.9515905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9515971Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9516203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9516270Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9516497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9516571Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9516799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9516868Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9516874Z 2025-09-07T08:14:18.9516947Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9517038Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9517221Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9517277Z return mod(**inputs) 2025-09-07T08:14:18.9517504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9517565Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9517794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9517855Z layer_outputs = layer_module( 2025-09-07T08:14:18.9518068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9518132Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9518358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9518446Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9518678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.9518770Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9518995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9519121Z return self.weight * hidden_states 2025-09-07T08:14:18.9519125Z 2025-09-07T08:14:18.9519215Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9519398Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9519457Z return mod(**inputs) 2025-09-07T08:14:18.9519689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9519787Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9520018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9520080Z layer_outputs = layer_module( 2025-09-07T08:14:18.9520298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9520368Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9520598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9520678Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9520906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9521014Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9521276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.9521359Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.9521363Z 2025-09-07T08:14:18.9521479Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9521667Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9521722Z return mod(**inputs) 2025-09-07T08:14:18.9521950Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9522016Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9522244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9522306Z layer_outputs = layer_module( 2025-09-07T08:14:18.9522518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9522585Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9522813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9522887Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9523114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9523216Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9523444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.9523515Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.9523519Z 2025-09-07T08:14:18.9523605Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9523790Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9523846Z return mod(**inputs) 2025-09-07T08:14:18.9524075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9524137Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9524364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9524425Z layer_outputs = layer_module( 2025-09-07T08:14:18.9524634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9524881Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9525107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9525181Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9525446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9525546Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9525781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.9525851Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.9525854Z 2025-09-07T08:14:18.9525930Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9526019Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9526204Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9526267Z return mod(**inputs) 2025-09-07T08:14:18.9526494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9526557Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9526819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9526880Z layer_outputs = layer_module( 2025-09-07T08:14:18.9527123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9527191Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9527418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9527490Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9527717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.9527817Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9528042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9528113Z return self.weight * hidden_states 2025-09-07T08:14:18.9528116Z 2025-09-07T08:14:18.9528203Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9528386Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9528441Z return mod(**inputs) 2025-09-07T08:14:18.9528668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9528731Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9528961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9529028Z layer_outputs = layer_module( 2025-09-07T08:14:18.9529243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9529310Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9529543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9529612Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9529842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9529912Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9530138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9530206Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9530243Z 2025-09-07T08:14:18.9530329Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9530511Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9530565Z return mod(**inputs) 2025-09-07T08:14:18.9530796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9530890Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9531118Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9531182Z layer_outputs = layer_module( 2025-09-07T08:14:18.9531392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9531459Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9531685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9531755Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9531982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9532050Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9532309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9532377Z key_states = self.k(current_states) 2025-09-07T08:14:18.9532381Z 2025-09-07T08:14:18.9532496Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9532679Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9532733Z return mod(**inputs) 2025-09-07T08:14:18.9532963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9533025Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9533256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9533315Z layer_outputs = layer_module( 2025-09-07T08:14:18.9533524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9533594Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9533818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9533889Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9534113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9534180Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9534414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9534537Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9534540Z 2025-09-07T08:14:18.9534630Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9534810Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9534869Z return mod(**inputs) 2025-09-07T08:14:18.9535101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9535163Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9535395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9535453Z layer_outputs = layer_module( 2025-09-07T08:14:18.9535666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9535763Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9535990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9536060Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9536289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9536395Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9536629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9536753Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9536757Z 2025-09-07T08:14:18.9536849Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9537037Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9537103Z return mod(**inputs) 2025-09-07T08:14:18.9537403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9537471Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9537701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9537801Z layer_outputs = layer_module( 2025-09-07T08:14:18.9538021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9538091Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9538374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9538446Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9538672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9538748Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9538972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9539086Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9539090Z 2025-09-07T08:14:18.9539180Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9539365Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9539421Z return mod(**inputs) 2025-09-07T08:14:18.9539650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9539714Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9539946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9540014Z layer_outputs = layer_module( 2025-09-07T08:14:18.9540227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9540296Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9540527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9540598Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9540830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9540903Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9541134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9541282Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9541286Z 2025-09-07T08:14:18.9541374Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9541591Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9541646Z return mod(**inputs) 2025-09-07T08:14:18.9541876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9541973Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9542203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9542266Z layer_outputs = layer_module( 2025-09-07T08:14:18.9542477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9542546Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9542777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9542848Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9543079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9543146Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9543407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9543477Z value_states = self.v(current_states) 2025-09-07T08:14:18.9543480Z 2025-09-07T08:14:18.9543570Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9543783Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9543839Z return mod(**inputs) 2025-09-07T08:14:18.9544069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9544134Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9544373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9544431Z layer_outputs = layer_module( 2025-09-07T08:14:18.9544640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9544710Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9544942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9545017Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9545242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9545309Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9545541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9545641Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9545645Z 2025-09-07T08:14:18.9545738Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9545919Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9545977Z return mod(**inputs) 2025-09-07T08:14:18.9546209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9546270Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9546500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9546559Z layer_outputs = layer_module( 2025-09-07T08:14:18.9546771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9546834Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9547097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9547171Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9547396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9547516Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9547744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9547840Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9547846Z 2025-09-07T08:14:18.9547941Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9548128Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9548187Z return mod(**inputs) 2025-09-07T08:14:18.9548421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9548488Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9548717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9548780Z layer_outputs = layer_module( 2025-09-07T08:14:18.9549037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9549106Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9549376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9549450Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9549684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9549766Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9549992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9550090Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9550094Z 2025-09-07T08:14:18.9550186Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9550378Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9550435Z return mod(**inputs) 2025-09-07T08:14:18.9550664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9550736Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9550965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9551030Z layer_outputs = layer_module( 2025-09-07T08:14:18.9551242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9551312Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9551541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9551610Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9551846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9551914Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9552140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9552235Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9552238Z 2025-09-07T08:14:18.9552327Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9552512Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9552601Z return mod(**inputs) 2025-09-07T08:14:18.9552833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9552895Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9553123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9553225Z layer_outputs = layer_module( 2025-09-07T08:14:18.9553436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9553510Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9553737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9553805Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9554035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9554108Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9554339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9554408Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9554413Z 2025-09-07T08:14:18.9554538Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9554723Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9554779Z return mod(**inputs) 2025-09-07T08:14:18.9555045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9555111Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9555343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9555409Z layer_outputs = layer_module( 2025-09-07T08:14:18.9555620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9555693Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9555922Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9555999Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9556227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:14:18.9556351Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:14:18.9556359Z 2025-09-07T08:14:18.9556434Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9556524Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9556709Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9556765Z return mod(**inputs) 2025-09-07T08:14:18.9556994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9557056Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9557283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9557347Z layer_outputs = layer_module( 2025-09-07T08:14:18.9557556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9557625Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9557851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9557920Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9558149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T08:14:18.9558280Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9558511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9558579Z return self.weight * hidden_states 2025-09-07T08:14:18.9558613Z 2025-09-07T08:14:18.9558704Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9558893Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9558951Z return mod(**inputs) 2025-09-07T08:14:18.9559187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9559252Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9559486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9559551Z layer_outputs = layer_module( 2025-09-07T08:14:18.9559763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9559840Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9560112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9560190Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9560418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9560524Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9560757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9560825Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9560828Z 2025-09-07T08:14:18.9560924Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9561105Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9561161Z return mod(**inputs) 2025-09-07T08:14:18.9561394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9561465Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9561695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9561756Z layer_outputs = layer_module( 2025-09-07T08:14:18.9561968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9562033Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9562257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9562329Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9562553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9562625Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9562850Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9562918Z key_states = self.k(current_states) 2025-09-07T08:14:18.9562921Z 2025-09-07T08:14:18.9563010Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9563192Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9563249Z return mod(**inputs) 2025-09-07T08:14:18.9563476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9563537Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9563809Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9563872Z layer_outputs = layer_module( 2025-09-07T08:14:18.9564084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9564185Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9564415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9564484Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9564710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9564782Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9565006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9565127Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9565130Z 2025-09-07T08:14:18.9565219Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9565399Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9565455Z return mod(**inputs) 2025-09-07T08:14:18.9565725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9565790Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9566051Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9566115Z layer_outputs = layer_module( 2025-09-07T08:14:18.9566326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9566390Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9566622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9566690Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9566916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9566989Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9567214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9567328Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9567332Z 2025-09-07T08:14:18.9567420Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9567602Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9567657Z return mod(**inputs) 2025-09-07T08:14:18.9567885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9567949Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9568175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9568237Z layer_outputs = layer_module( 2025-09-07T08:14:18.9568450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9568517Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9568743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9568811Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9569038Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9569109Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9569382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9569489Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9569492Z 2025-09-07T08:14:18.9569581Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9569804Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9569860Z return mod(**inputs) 2025-09-07T08:14:18.9570092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9570153Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9570381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9570443Z layer_outputs = layer_module( 2025-09-07T08:14:18.9570655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9570721Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9570946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9571017Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9571277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9571353Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9571611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9571755Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9571759Z 2025-09-07T08:14:18.9571848Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9572031Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9572085Z return mod(**inputs) 2025-09-07T08:14:18.9572314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9572375Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9572608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9572667Z layer_outputs = layer_module( 2025-09-07T08:14:18.9572879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9572944Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9573173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9573245Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9573471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9573543Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9573769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9573839Z value_states = self.v(current_states) 2025-09-07T08:14:18.9573843Z 2025-09-07T08:14:18.9573931Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9574113Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9574170Z return mod(**inputs) 2025-09-07T08:14:18.9574397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9574458Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9574688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9574787Z layer_outputs = layer_module( 2025-09-07T08:14:18.9575000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9575065Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9575328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9575397Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9575623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9575696Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9575921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9576019Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9576023Z 2025-09-07T08:14:18.9576111Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9576293Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9576349Z return mod(**inputs) 2025-09-07T08:14:18.9576611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9576677Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9576938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9577002Z layer_outputs = layer_module( 2025-09-07T08:14:18.9577272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9577338Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9577565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9577635Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9577862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9577931Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9578157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9578251Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9578255Z 2025-09-07T08:14:18.9578344Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9578526Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9578581Z return mod(**inputs) 2025-09-07T08:14:18.9578809Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9578873Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9579101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9579162Z layer_outputs = layer_module( 2025-09-07T08:14:18.9579374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9579441Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9579669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9579737Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9579966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9580036Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9580262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9580388Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9580391Z 2025-09-07T08:14:18.9580478Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9580663Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9580767Z return mod(**inputs) 2025-09-07T08:14:18.9581001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9581065Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9581292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9581354Z layer_outputs = layer_module( 2025-09-07T08:14:18.9581572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9581644Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9581876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9581945Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9582218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9582291Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9582550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9582640Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9582643Z 2025-09-07T08:14:18.9582733Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9582913Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9582969Z return mod(**inputs) 2025-09-07T08:14:18.9583200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9583261Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9583494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9583555Z layer_outputs = layer_module( 2025-09-07T08:14:18.9583765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9583833Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9584058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9584129Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9584352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9584425Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9584650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9584715Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9584719Z 2025-09-07T08:14:18.9584793Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9584880Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9585060Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9585116Z return mod(**inputs) 2025-09-07T08:14:18.9585344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9585406Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9585633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9585731Z layer_outputs = layer_module( 2025-09-07T08:14:18.9585941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9586006Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9586238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9586353Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9586583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.9586668Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9586894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9586959Z return self.weight * hidden_states 2025-09-07T08:14:18.9586963Z 2025-09-07T08:14:18.9587050Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9587235Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9587290Z return mod(**inputs) 2025-09-07T08:14:18.9587519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9587618Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9587849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9587984Z layer_outputs = layer_module( 2025-09-07T08:14:18.9588201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9588270Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9588494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9588575Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9588801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9588909Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9589137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.9589206Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.9589210Z 2025-09-07T08:14:18.9589300Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9589481Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9589536Z return mod(**inputs) 2025-09-07T08:14:18.9589768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9589830Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9590063Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9590122Z layer_outputs = layer_module( 2025-09-07T08:14:18.9590331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9590402Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9590628Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9590706Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9590929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9591032Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9591258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.9591366Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.9591369Z 2025-09-07T08:14:18.9591458Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9591637Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9591729Z return mod(**inputs) 2025-09-07T08:14:18.9591961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9592022Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9592251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9592312Z layer_outputs = layer_module( 2025-09-07T08:14:18.9592524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9592592Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9592816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9592893Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9593149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9593253Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9593515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.9593586Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.9593589Z 2025-09-07T08:14:18.9593657Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9593745Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9593927Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9593983Z return mod(**inputs) 2025-09-07T08:14:18.9594213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9594272Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9594500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9594564Z layer_outputs = layer_module( 2025-09-07T08:14:18.9594774Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9594842Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9595066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9595135Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9595362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.9595457Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9595684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9595749Z return self.weight * hidden_states 2025-09-07T08:14:18.9595753Z 2025-09-07T08:14:18.9595844Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9596024Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9596079Z return mod(**inputs) 2025-09-07T08:14:18.9596311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9596370Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9596601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9596694Z layer_outputs = layer_module( 2025-09-07T08:14:18.9596910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9596978Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9597207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9597311Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9597537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9597611Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9597836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9597903Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9597906Z 2025-09-07T08:14:18.9597997Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9598177Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9598233Z return mod(**inputs) 2025-09-07T08:14:18.9598460Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9598556Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9598789Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9598849Z layer_outputs = layer_module( 2025-09-07T08:14:18.9599090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9599155Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9599380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9599453Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9599678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9599750Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9599974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9600044Z key_states = self.k(current_states) 2025-09-07T08:14:18.9600047Z 2025-09-07T08:14:18.9600134Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9600316Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9600374Z return mod(**inputs) 2025-09-07T08:14:18.9600602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9600665Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9600895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9600954Z layer_outputs = layer_module( 2025-09-07T08:14:18.9601174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9601241Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9601469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9601536Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9601763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9601840Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9602066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9602253Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9602256Z 2025-09-07T08:14:18.9602346Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9602530Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9602585Z return mod(**inputs) 2025-09-07T08:14:18.9602855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9602925Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9603303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9603366Z layer_outputs = layer_module( 2025-09-07T08:14:18.9603581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9603651Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9603881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9603952Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9604184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9604325Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9604558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9604721Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9604724Z 2025-09-07T08:14:18.9604817Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9605001Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9605057Z return mod(**inputs) 2025-09-07T08:14:18.9605294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9605356Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9605584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9605652Z layer_outputs = layer_module( 2025-09-07T08:14:18.9605864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9605935Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9606164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9606237Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9606463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9606534Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9606767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9606876Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9606879Z 2025-09-07T08:14:18.9606969Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9607154Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9607212Z return mod(**inputs) 2025-09-07T08:14:18.9607449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9607511Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9607742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9607805Z layer_outputs = layer_module( 2025-09-07T08:14:18.9608074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9608146Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9608378Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9608451Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9608727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9608800Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9609028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9609173Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9609176Z 2025-09-07T08:14:18.9609271Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9609459Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9609517Z return mod(**inputs) 2025-09-07T08:14:18.9609748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9609811Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9610080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9610143Z layer_outputs = layer_module( 2025-09-07T08:14:18.9610392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9610461Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9610691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9610763Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9610991Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9611062Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9611286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9611360Z value_states = self.v(current_states) 2025-09-07T08:14:18.9611363Z 2025-09-07T08:14:18.9611451Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9611635Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9611694Z return mod(**inputs) 2025-09-07T08:14:18.9611923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9611985Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9612211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9612273Z layer_outputs = layer_module( 2025-09-07T08:14:18.9612488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9612552Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9612783Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9612850Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9613075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9613146Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9613374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9613472Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9613513Z 2025-09-07T08:14:18.9613602Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9613786Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9613841Z return mod(**inputs) 2025-09-07T08:14:18.9614077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9614173Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9614404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9614470Z layer_outputs = layer_module( 2025-09-07T08:14:18.9614683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9614751Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9614984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9615054Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9615281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9615349Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9615605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9615702Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9615706Z 2025-09-07T08:14:18.9615822Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9616008Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9616062Z return mod(**inputs) 2025-09-07T08:14:18.9616304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9616366Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9616594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9616656Z layer_outputs = layer_module( 2025-09-07T08:14:18.9616875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9616946Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9617218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9617288Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9617519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9617589Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9617816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9617907Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9617911Z 2025-09-07T08:14:18.9618003Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9618189Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9618248Z return mod(**inputs) 2025-09-07T08:14:18.9618478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9618542Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9618773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9618833Z layer_outputs = layer_module( 2025-09-07T08:14:18.9619044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9619157Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9619385Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9619456Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9619682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9619792Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9620025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9620116Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9620119Z 2025-09-07T08:14:18.9620211Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9620392Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9620450Z return mod(**inputs) 2025-09-07T08:14:18.9620679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9620739Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9620971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9621067Z layer_outputs = layer_module( 2025-09-07T08:14:18.9621282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9621399Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9621629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9621704Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9621930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9622003Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9622227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9622293Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9622299Z 2025-09-07T08:14:18.9622370Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9622459Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9622643Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9622697Z return mod(**inputs) 2025-09-07T08:14:18.9622929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9622989Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9623218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9623282Z layer_outputs = layer_module( 2025-09-07T08:14:18.9623492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9623559Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9623785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9623855Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9624081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T08:14:18.9624176Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9624406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9624470Z return self.weight * hidden_states 2025-09-07T08:14:18.9624473Z 2025-09-07T08:14:18.9624613Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9624795Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9624849Z return mod(**inputs) 2025-09-07T08:14:18.9625081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9625178Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9625414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9625475Z layer_outputs = layer_module( 2025-09-07T08:14:18.9625687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9625753Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9625978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9626058Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9626282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9626354Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9626616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9626685Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9626689Z 2025-09-07T08:14:18.9626779Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9626996Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9627057Z return mod(**inputs) 2025-09-07T08:14:18.9627290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9627351Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9627595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9627655Z layer_outputs = layer_module( 2025-09-07T08:14:18.9627872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9627943Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9628170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9628248Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9628475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9628549Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9628774Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9628843Z key_states = self.k(current_states) 2025-09-07T08:14:18.9628849Z 2025-09-07T08:14:18.9628937Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9629116Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9629172Z return mod(**inputs) 2025-09-07T08:14:18.9629402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9629466Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9629695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9629754Z layer_outputs = layer_module( 2025-09-07T08:14:18.9629965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9630029Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9630300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9630369Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9630594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9630703Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9630930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9631052Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9631055Z 2025-09-07T08:14:18.9631143Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9631324Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9631381Z return mod(**inputs) 2025-09-07T08:14:18.9631608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9631676Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9631904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9631968Z layer_outputs = layer_module( 2025-09-07T08:14:18.9632214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9632280Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9632539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9632610Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9632838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9632908Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9633136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9633247Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9633250Z 2025-09-07T08:14:18.9633338Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9633524Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9633579Z return mod(**inputs) 2025-09-07T08:14:18.9633813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9633876Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9634103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9634167Z layer_outputs = layer_module( 2025-09-07T08:14:18.9634381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9634452Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9634678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9634746Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9634975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9635046Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9635275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9635381Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9635384Z 2025-09-07T08:14:18.9635474Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9635657Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9635753Z return mod(**inputs) 2025-09-07T08:14:18.9635982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9636044Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9636313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9636374Z layer_outputs = layer_module( 2025-09-07T08:14:18.9636585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9636653Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9636879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9636949Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9637176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9637246Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9637472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9637652Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9637656Z 2025-09-07T08:14:18.9637746Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9637965Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9638023Z return mod(**inputs) 2025-09-07T08:14:18.9638252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9638313Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9638542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9638604Z layer_outputs = layer_module( 2025-09-07T08:14:18.9638814Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9638878Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9639106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9639176Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9639401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9639474Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9639697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9639765Z value_states = self.v(current_states) 2025-09-07T08:14:18.9639769Z 2025-09-07T08:14:18.9639862Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9640042Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9640100Z return mod(**inputs) 2025-09-07T08:14:18.9640328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9640393Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9640620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9640681Z layer_outputs = layer_module( 2025-09-07T08:14:18.9640894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9640958Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9641186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9641289Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9641515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9641587Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9641850Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9641948Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9641952Z 2025-09-07T08:14:18.9642039Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9642219Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9642277Z return mod(**inputs) 2025-09-07T08:14:18.9642503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9642568Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9642795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9642857Z layer_outputs = layer_module( 2025-09-07T08:14:18.9643113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9643180Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9643439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9643508Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9643738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9643815Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9644043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9644139Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9644143Z 2025-09-07T08:14:18.9644229Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9644410Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9644467Z return mod(**inputs) 2025-09-07T08:14:18.9644695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9644759Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9644986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9645050Z layer_outputs = layer_module( 2025-09-07T08:14:18.9645259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9645327Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9645550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9645618Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9645845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9645917Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9646147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9646235Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9646238Z 2025-09-07T08:14:18.9646325Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9646507Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9646596Z return mod(**inputs) 2025-09-07T08:14:18.9646827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9646888Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9647118Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9647214Z layer_outputs = layer_module( 2025-09-07T08:14:18.9647425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9647494Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9647721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9647793Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9648018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9648089Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9648315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9648404Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9648413Z 2025-09-07T08:14:18.9648533Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9648713Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9648768Z return mod(**inputs) 2025-09-07T08:14:18.9649029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9649092Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9649322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9649384Z layer_outputs = layer_module( 2025-09-07T08:14:18.9649597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9649661Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9649886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9649961Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9650184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9650257Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9650480Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9650544Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9650548Z 2025-09-07T08:14:18.9650638Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9650822Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9650879Z return mod(**inputs) 2025-09-07T08:14:18.9651108Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9651172Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9651405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9651465Z layer_outputs = layer_module( 2025-09-07T08:14:18.9651680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9651745Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9651971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9652039Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9652302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 647, in forward 2025-09-07T08:14:18.9652432Z layer_output = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:14:18.9652435Z 2025-09-07T08:14:18.9652508Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9652638Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9652821Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9652877Z return mod(**inputs) 2025-09-07T08:14:18.9653115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9653178Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9653410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9653473Z layer_outputs = layer_module( 2025-09-07T08:14:18.9653685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9653757Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9653987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9654109Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9654339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.9654463Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9654695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9654765Z return self.weight * hidden_states 2025-09-07T08:14:18.9654768Z 2025-09-07T08:14:18.9654863Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9655046Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9655109Z return mod(**inputs) 2025-09-07T08:14:18.9655340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9655408Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9655640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9655702Z layer_outputs = layer_module( 2025-09-07T08:14:18.9655913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9655982Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9656211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9656289Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9656517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9656629Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9656854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.9656927Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.9656930Z 2025-09-07T08:14:18.9657019Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9657250Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9657308Z return mod(**inputs) 2025-09-07T08:14:18.9657537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9657602Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9657872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9657934Z layer_outputs = layer_module( 2025-09-07T08:14:18.9658145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9658248Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9658476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9658550Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9658778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9658878Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9659103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.9659177Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.9659180Z 2025-09-07T08:14:18.9659266Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9659447Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9659502Z return mod(**inputs) 2025-09-07T08:14:18.9659763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9659827Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9660088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9660153Z layer_outputs = layer_module( 2025-09-07T08:14:18.9660363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9660432Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9660665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9660738Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9660965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9661065Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9661291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.9661359Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.9661362Z 2025-09-07T08:14:18.9661431Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9661521Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9661703Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9661761Z return mod(**inputs) 2025-09-07T08:14:18.9661989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9662050Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9662281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9662347Z layer_outputs = layer_module( 2025-09-07T08:14:18.9662560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9662624Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9662850Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9662922Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9663146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.9663275Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9663499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9663567Z return self.weight * hidden_states 2025-09-07T08:14:18.9663570Z 2025-09-07T08:14:18.9663705Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9663887Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9663945Z return mod(**inputs) 2025-09-07T08:14:18.9664175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9664239Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9664467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9664528Z layer_outputs = layer_module( 2025-09-07T08:14:18.9664746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9664812Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9665039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9665143Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9665374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9665476Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9665705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9665774Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9665778Z 2025-09-07T08:14:18.9665869Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9666058Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9666114Z return mod(**inputs) 2025-09-07T08:14:18.9666355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9666424Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9666655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9666725Z layer_outputs = layer_module( 2025-09-07T08:14:18.9666935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9667004Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9667232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9671560Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9671866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9671948Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9672200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9672279Z key_states = self.k(current_states) 2025-09-07T08:14:18.9672284Z 2025-09-07T08:14:18.9672387Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9672599Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9672663Z return mod(**inputs) 2025-09-07T08:14:18.9672914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9672989Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9673232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9673368Z layer_outputs = layer_module( 2025-09-07T08:14:18.9673603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9673680Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9673968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9674045Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9674288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9674363Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9674596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9674719Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9674725Z 2025-09-07T08:14:18.9674827Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9675030Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9675090Z return mod(**inputs) 2025-09-07T08:14:18.9675380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9675449Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9675713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9675784Z layer_outputs = layer_module( 2025-09-07T08:14:18.9676003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9676085Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9676319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9676398Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9676624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9676698Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9676929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9677043Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9677047Z 2025-09-07T08:14:18.9677147Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9677333Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9677394Z return mod(**inputs) 2025-09-07T08:14:18.9677644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9677714Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9677965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9678026Z layer_outputs = layer_module( 2025-09-07T08:14:18.9678246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9678319Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9678552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9678628Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9678854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9678936Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9679203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9679319Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9679323Z 2025-09-07T08:14:18.9679423Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9679618Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9679725Z return mod(**inputs) 2025-09-07T08:14:18.9679970Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9680039Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9680284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9680346Z layer_outputs = layer_module( 2025-09-07T08:14:18.9680567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9680638Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9680870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9680944Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9681246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9681324Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9681585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9681738Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9681742Z 2025-09-07T08:14:18.9681837Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9682033Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9682100Z return mod(**inputs) 2025-09-07T08:14:18.9682334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9682406Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9682639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9682704Z layer_outputs = layer_module( 2025-09-07T08:14:18.9682923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9682993Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9683229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9683302Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9683536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9683616Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9683844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9683924Z value_states = self.v(current_states) 2025-09-07T08:14:18.9683929Z 2025-09-07T08:14:18.9684025Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9684214Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9684273Z return mod(**inputs) 2025-09-07T08:14:18.9684507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9684575Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9684811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9684937Z layer_outputs = layer_module( 2025-09-07T08:14:18.9685157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9685236Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9685473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9685578Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9685811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9685884Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9686117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9686219Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9686222Z 2025-09-07T08:14:18.9686321Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9686518Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9686580Z return mod(**inputs) 2025-09-07T08:14:18.9686819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9686920Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9687158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9687224Z layer_outputs = layer_module( 2025-09-07T08:14:18.9687468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9687540Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9687768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9687841Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9688065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9688136Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9688364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9688456Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9688459Z 2025-09-07T08:14:18.9688551Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9688735Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9688791Z return mod(**inputs) 2025-09-07T08:14:18.9689022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9689084Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9689318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9689380Z layer_outputs = layer_module( 2025-09-07T08:14:18.9689592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9689664Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9689893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9689966Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9690191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9690262Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9690486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9690627Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9690631Z 2025-09-07T08:14:18.9690721Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9690905Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9690998Z return mod(**inputs) 2025-09-07T08:14:18.9691233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9691301Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9691540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9691600Z layer_outputs = layer_module( 2025-09-07T08:14:18.9691812Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9691878Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9692112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9692180Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9692406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9692514Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9692748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9692882Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9692885Z 2025-09-07T08:14:18.9692975Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9693159Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9693223Z return mod(**inputs) 2025-09-07T08:14:18.9693456Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9693521Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9693752Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9693815Z layer_outputs = layer_module( 2025-09-07T08:14:18.9694036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9694105Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9694340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9694413Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9694644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9694716Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9694949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9695024Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9695028Z 2025-09-07T08:14:18.9695101Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9695201Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9695385Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9695443Z return mod(**inputs) 2025-09-07T08:14:18.9695682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9695748Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9695986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9696046Z layer_outputs = layer_module( 2025-09-07T08:14:18.9696302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9696380Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9696606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9696720Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9696949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T08:14:18.9697056Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9697354Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9697424Z return self.weight * hidden_states 2025-09-07T08:14:18.9697427Z 2025-09-07T08:14:18.9697529Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9697714Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9697778Z return mod(**inputs) 2025-09-07T08:14:18.9698015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9698079Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9698356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9698422Z layer_outputs = layer_module( 2025-09-07T08:14:18.9698678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9698749Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9698982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9699059Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9699288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9699373Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9699602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9699681Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9699684Z 2025-09-07T08:14:18.9699777Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9699962Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9700026Z return mod(**inputs) 2025-09-07T08:14:18.9700256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9700322Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9700549Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9700611Z layer_outputs = layer_module( 2025-09-07T08:14:18.9700828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9700893Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9701126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9701196Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9701422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9701497Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9701720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9701792Z key_states = self.k(current_states) 2025-09-07T08:14:18.9701827Z 2025-09-07T08:14:18.9701917Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9702101Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9702155Z return mod(**inputs) 2025-09-07T08:14:18.9702386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9702483Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9702713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9702776Z layer_outputs = layer_module( 2025-09-07T08:14:18.9702986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9703220Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9703455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9703526Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9703763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9703835Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9704151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9704275Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9704278Z 2025-09-07T08:14:18.9704415Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9704603Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9704658Z return mod(**inputs) 2025-09-07T08:14:18.9704890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9704954Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9705185Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9705252Z layer_outputs = layer_module( 2025-09-07T08:14:18.9705463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9705532Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9705758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9705828Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9706058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9706128Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9706360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9706469Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9706473Z 2025-09-07T08:14:18.9706566Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9706750Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9706808Z return mod(**inputs) 2025-09-07T08:14:18.9707041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9707102Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9707335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9707395Z layer_outputs = layer_module( 2025-09-07T08:14:18.9707606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9707723Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9707951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9708025Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9708302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9708377Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9708609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9708723Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9708726Z 2025-09-07T08:14:18.9708825Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9709009Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9709073Z return mod(**inputs) 2025-09-07T08:14:18.9709307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9709369Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9709636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9709701Z layer_outputs = layer_module( 2025-09-07T08:14:18.9709920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9710039Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9710272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9710345Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9710578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9710658Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9710882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9711031Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9711037Z 2025-09-07T08:14:18.9711127Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9711312Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9711373Z return mod(**inputs) 2025-09-07T08:14:18.9711603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9711667Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9711893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9711955Z layer_outputs = layer_module( 2025-09-07T08:14:18.9712169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9712233Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9712464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9712534Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9712761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9712835Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9713058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9713128Z value_states = self.v(current_states) 2025-09-07T08:14:18.9713132Z 2025-09-07T08:14:18.9713254Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9713437Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9713493Z return mod(**inputs) 2025-09-07T08:14:18.9713721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9713820Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9714051Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9714115Z layer_outputs = layer_module( 2025-09-07T08:14:18.9714325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9714390Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9714621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9714690Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9714916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9714986Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9715241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9715345Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9715348Z 2025-09-07T08:14:18.9715435Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9715650Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9715706Z return mod(**inputs) 2025-09-07T08:14:18.9715938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9716001Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9716230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9716293Z layer_outputs = layer_module( 2025-09-07T08:14:18.9716504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9716574Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9716802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9716872Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9717099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9717170Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9717398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9717488Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9717491Z 2025-09-07T08:14:18.9717581Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9717763Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9717820Z return mod(**inputs) 2025-09-07T08:14:18.9718053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9718114Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9718343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9718402Z layer_outputs = layer_module( 2025-09-07T08:14:18.9718612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9718679Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9718940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9719012Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9719238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9719346Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9719582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9719671Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9719674Z 2025-09-07T08:14:18.9719767Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9719946Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9720003Z return mod(**inputs) 2025-09-07T08:14:18.9720231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9720293Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9720524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9720587Z layer_outputs = layer_module( 2025-09-07T08:14:18.9720836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9720901Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9721157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9721234Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9721458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9721533Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9721757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9721852Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9721855Z 2025-09-07T08:14:18.9721942Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9722130Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9722192Z return mod(**inputs) 2025-09-07T08:14:18.9722422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9722486Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9722714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9722773Z layer_outputs = layer_module( 2025-09-07T08:14:18.9722990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9723055Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9723289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9723359Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9723585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9723662Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9723886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9723959Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9723962Z 2025-09-07T08:14:18.9724034Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9724130Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9724355Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9724419Z return mod(**inputs) 2025-09-07T08:14:18.9724659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9724762Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9724998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9725061Z layer_outputs = layer_module( 2025-09-07T08:14:18.9725272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9725339Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9725565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9725650Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9725877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.9725968Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9726232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9726302Z return self.weight * hidden_states 2025-09-07T08:14:18.9726305Z 2025-09-07T08:14:18.9726396Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9726610Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9726669Z return mod(**inputs) 2025-09-07T08:14:18.9726898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9726960Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9727192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9727252Z layer_outputs = layer_module( 2025-09-07T08:14:18.9727465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9727532Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9727760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9727836Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9728064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9728175Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9728400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.9728476Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.9728480Z 2025-09-07T08:14:18.9728568Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9728751Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9728809Z return mod(**inputs) 2025-09-07T08:14:18.9729045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9729110Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9729342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9729402Z layer_outputs = layer_module( 2025-09-07T08:14:18.9729615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9729681Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9729963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9730039Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9730271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9730415Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9730642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.9730717Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.9730721Z 2025-09-07T08:14:18.9730808Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9730991Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9731046Z return mod(**inputs) 2025-09-07T08:14:18.9731274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9731339Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9731569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9731632Z layer_outputs = layer_module( 2025-09-07T08:14:18.9731897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9731966Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9732225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9732300Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9732529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9732628Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9732856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.9732924Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.9732927Z 2025-09-07T08:14:18.9733014Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9733200Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9733255Z return mod(**inputs) 2025-09-07T08:14:18.9733485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9733546Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9733773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9733834Z layer_outputs = layer_module( 2025-09-07T08:14:18.9734043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9734113Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9734337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9734413Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9734640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 343, in forward 2025-09-07T08:14:18.9734757Z hidden_states = hidden_states + self.dropout(forwarded_states) 2025-09-07T08:14:18.9734762Z 2025-09-07T08:14:18.9734833Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9734921Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9735105Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9735159Z return mod(**inputs) 2025-09-07T08:14:18.9735425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9735492Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9735720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9735822Z layer_outputs = layer_module( 2025-09-07T08:14:18.9736034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9736099Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9736328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9736399Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9736627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.9736727Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9736957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9737023Z return self.weight * hidden_states 2025-09-07T08:14:18.9737026Z 2025-09-07T08:14:18.9737116Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9737396Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9737454Z return mod(**inputs) 2025-09-07T08:14:18.9737721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9737785Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9738022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9738087Z layer_outputs = layer_module( 2025-09-07T08:14:18.9738301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9738372Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9738601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9738677Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9738908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9738986Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9739214Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9739281Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9739284Z 2025-09-07T08:14:18.9739374Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9739556Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9739613Z return mod(**inputs) 2025-09-07T08:14:18.9739842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9739902Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9740135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9740196Z layer_outputs = layer_module( 2025-09-07T08:14:18.9740406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9740473Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9740699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9740770Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9740993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9741100Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9741326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9741392Z key_states = self.k(current_states) 2025-09-07T08:14:18.9741628Z 2025-09-07T08:14:18.9741719Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9741900Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9741959Z return mod(**inputs) 2025-09-07T08:14:18.9742188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9742249Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9742481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9742542Z layer_outputs = layer_module( 2025-09-07T08:14:18.9742753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9742818Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9743075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9743149Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9743406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9743480Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9743704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9743824Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9743829Z 2025-09-07T08:14:18.9743917Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9744096Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9744154Z return mod(**inputs) 2025-09-07T08:14:18.9744381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9744448Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9744674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9744734Z layer_outputs = layer_module( 2025-09-07T08:14:18.9744946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9745011Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9745240Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9745308Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9745531Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9745602Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9745827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9745940Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9745943Z 2025-09-07T08:14:18.9746031Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9746215Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9746270Z return mod(**inputs) 2025-09-07T08:14:18.9746497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9746595Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9746822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9746884Z layer_outputs = layer_module( 2025-09-07T08:14:18.9747095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9747194Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9747422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9747492Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9747719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9747787Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9748013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9748122Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9748126Z 2025-09-07T08:14:18.9748213Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9748396Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9748485Z return mod(**inputs) 2025-09-07T08:14:18.9748722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9748784Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9749045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9749108Z layer_outputs = layer_module( 2025-09-07T08:14:18.9749319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9749389Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9749613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9749680Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9749908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9749978Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9750204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9750349Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9750353Z 2025-09-07T08:14:18.9750443Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9750625Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9750682Z return mod(**inputs) 2025-09-07T08:14:18.9750914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9750974Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9751204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9751267Z layer_outputs = layer_module( 2025-09-07T08:14:18.9751477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9751548Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9751774Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9751844Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9752067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9752174Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9752400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9752467Z value_states = self.v(current_states) 2025-09-07T08:14:18.9752470Z 2025-09-07T08:14:18.9752609Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9752790Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9752848Z return mod(**inputs) 2025-09-07T08:14:18.9753079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9753140Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9753373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9753432Z layer_outputs = layer_module( 2025-09-07T08:14:18.9753646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9753711Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9753936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9754040Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9754267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9754369Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9754596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9754697Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9754700Z 2025-09-07T08:14:18.9754791Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9754973Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9755030Z return mod(**inputs) 2025-09-07T08:14:18.9755258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9755320Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9755551Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9755613Z layer_outputs = layer_module( 2025-09-07T08:14:18.9755829Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9755894Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9756122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9756191Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9756416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9756488Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9756711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9756807Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9756810Z 2025-09-07T08:14:18.9756898Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9757081Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9757136Z return mod(**inputs) 2025-09-07T08:14:18.9757382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9757444Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9757709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9757772Z layer_outputs = layer_module( 2025-09-07T08:14:18.9757983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9758048Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9758311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9758379Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9758611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9758679Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9758904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9758998Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9759002Z 2025-09-07T08:14:18.9759087Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9759270Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9759324Z return mod(**inputs) 2025-09-07T08:14:18.9759591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9759653Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9759911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9759974Z layer_outputs = layer_module( 2025-09-07T08:14:18.9760183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9760251Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9760479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9760548Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9760775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9760846Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9761075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9761169Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9761173Z 2025-09-07T08:14:18.9761263Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9761443Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9761498Z return mod(**inputs) 2025-09-07T08:14:18.9761728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9761790Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9762022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9762081Z layer_outputs = layer_module( 2025-09-07T08:14:18.9762294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9762363Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9762590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9762660Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9762883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9762951Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9763216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9763284Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9763287Z 2025-09-07T08:14:18.9763361Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9763449Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9763670Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9763726Z return mod(**inputs) 2025-09-07T08:14:18.9763959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9764022Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9764253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9764314Z layer_outputs = layer_module( 2025-09-07T08:14:18.9764524Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9764591Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9764818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9764887Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9765146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T08:14:18.9765243Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9765508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9765579Z return self.weight * hidden_states 2025-09-07T08:14:18.9765582Z 2025-09-07T08:14:18.9765671Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9765858Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9765916Z return mod(**inputs) 2025-09-07T08:14:18.9766147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9766208Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9766443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9766507Z layer_outputs = layer_module( 2025-09-07T08:14:18.9766717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9766785Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9767013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9767084Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9767313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9767385Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9767612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9767681Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9767686Z 2025-09-07T08:14:18.9767777Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9767958Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9768013Z return mod(**inputs) 2025-09-07T08:14:18.9768245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9768306Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9768536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9768632Z layer_outputs = layer_module( 2025-09-07T08:14:18.9768844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9768912Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9769138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9769270Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9769497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9769569Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9769795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9769860Z key_states = self.k(current_states) 2025-09-07T08:14:18.9769863Z 2025-09-07T08:14:18.9769953Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9770133Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9770191Z return mod(**inputs) 2025-09-07T08:14:18.9770421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9770518Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9770752Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9770844Z layer_outputs = layer_module( 2025-09-07T08:14:18.9771061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9771126Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9771351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9771424Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9771651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9771723Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9771948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9772068Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9772074Z 2025-09-07T08:14:18.9772164Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9772345Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9772401Z return mod(**inputs) 2025-09-07T08:14:18.9772629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9772697Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9772925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9772986Z layer_outputs = layer_module( 2025-09-07T08:14:18.9773199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9773266Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9773494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9773564Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9773788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9773861Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9774085Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9774256Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9774259Z 2025-09-07T08:14:18.9774346Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9774530Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9774619Z return mod(**inputs) 2025-09-07T08:14:18.9774849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9774914Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9775143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9775207Z layer_outputs = layer_module( 2025-09-07T08:14:18.9775418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9775487Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9775715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9775783Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9776053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9776127Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9776352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9776492Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9776496Z 2025-09-07T08:14:18.9776584Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9776766Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9776824Z return mod(**inputs) 2025-09-07T08:14:18.9777069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9777132Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9777421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9777494Z layer_outputs = layer_module( 2025-09-07T08:14:18.9777709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9777784Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9778016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9778089Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9778320Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9778393Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9778627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9778773Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9778778Z 2025-09-07T08:14:18.9778878Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9779068Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9779126Z return mod(**inputs) 2025-09-07T08:14:18.9779365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9779427Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9779663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9779767Z layer_outputs = layer_module( 2025-09-07T08:14:18.9779982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9780054Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9780282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9780406Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9780632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9780709Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9780939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9781009Z value_states = self.v(current_states) 2025-09-07T08:14:18.9781012Z 2025-09-07T08:14:18.9781107Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9781290Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9781352Z return mod(**inputs) 2025-09-07T08:14:18.9781585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9781651Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9781923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9781986Z layer_outputs = layer_module( 2025-09-07T08:14:18.9782247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9782314Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9782545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9782618Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9782848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9782926Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9783156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9783258Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9783265Z 2025-09-07T08:14:18.9783358Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9783544Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9783609Z return mod(**inputs) 2025-09-07T08:14:18.9783840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9783906Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9784138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9784199Z layer_outputs = layer_module( 2025-09-07T08:14:18.9784415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9784483Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9784718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9784788Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9785017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9785096Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9785321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9785457Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9785460Z 2025-09-07T08:14:18.9785549Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9785732Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9785793Z return mod(**inputs) 2025-09-07T08:14:18.9786060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9786130Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9786360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9786428Z layer_outputs = layer_module( 2025-09-07T08:14:18.9786641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9786708Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9786940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9787011Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9787244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9787351Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9787582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9787682Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9787716Z 2025-09-07T08:14:18.9787806Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9787993Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9788048Z return mod(**inputs) 2025-09-07T08:14:18.9788282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9788353Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9788582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9788648Z layer_outputs = layer_module( 2025-09-07T08:14:18.9788866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9788938Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9789164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9789236Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9789467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9789539Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9789773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9789862Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9789865Z 2025-09-07T08:14:18.9789953Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9790141Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9790197Z return mod(**inputs) 2025-09-07T08:14:18.9790437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9790498Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9790733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9790795Z layer_outputs = layer_module( 2025-09-07T08:14:18.9791013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9791120Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9791350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9791426Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9791689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9791763Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9791997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9792064Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9792067Z 2025-09-07T08:14:18.9792147Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9792236Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9792421Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9792482Z return mod(**inputs) 2025-09-07T08:14:18.9792712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9792779Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9793048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9793117Z layer_outputs = layer_module( 2025-09-07T08:14:18.9793383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9793451Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9793685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9793767Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9794003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.9794090Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9794317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9794395Z return self.weight * hidden_states 2025-09-07T08:14:18.9794398Z 2025-09-07T08:14:18.9794487Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9794674Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9794730Z return mod(**inputs) 2025-09-07T08:14:18.9794961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9795030Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9795259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9795326Z layer_outputs = layer_module( 2025-09-07T08:14:18.9795535Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9795608Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9795840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9795919Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9796151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9796260Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9796493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.9796565Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.9796605Z 2025-09-07T08:14:18.9796694Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9796883Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9796941Z return mod(**inputs) 2025-09-07T08:14:18.9797179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9797275Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9797509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9797577Z layer_outputs = layer_module( 2025-09-07T08:14:18.9797789Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9797860Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9798088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9798173Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9798399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9798500Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9798767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.9798844Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.9798847Z 2025-09-07T08:14:18.9798971Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9799157Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9799217Z return mod(**inputs) 2025-09-07T08:14:18.9799453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9799519Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9799752Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9799813Z layer_outputs = layer_module( 2025-09-07T08:14:18.9800033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9800102Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9800330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9800411Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9800640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9800746Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9800974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.9801042Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.9801045Z 2025-09-07T08:14:18.9801122Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9801210Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9801402Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9801458Z return mod(**inputs) 2025-09-07T08:14:18.9801687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9801751Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9801978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9802039Z layer_outputs = layer_module( 2025-09-07T08:14:18.9802288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9802357Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9802582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9802653Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9802918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.9803014Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9803397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9803465Z return self.weight * hidden_states 2025-09-07T08:14:18.9803468Z 2025-09-07T08:14:18.9803557Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9803742Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9803800Z return mod(**inputs) 2025-09-07T08:14:18.9804031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9804093Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9804410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9804474Z layer_outputs = layer_module( 2025-09-07T08:14:18.9804727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9804796Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9805022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9805094Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9805320Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9805392Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9805619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9805687Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9805692Z 2025-09-07T08:14:18.9805784Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9805964Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9806020Z return mod(**inputs) 2025-09-07T08:14:18.9806250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9806310Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9806541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9806602Z layer_outputs = layer_module( 2025-09-07T08:14:18.9806812Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9806879Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9807106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9807177Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9807402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9807473Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9807697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9807762Z key_states = self.k(current_states) 2025-09-07T08:14:18.9807766Z 2025-09-07T08:14:18.9807904Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9808085Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9808142Z return mod(**inputs) 2025-09-07T08:14:18.9808368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9808478Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9808711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9808773Z layer_outputs = layer_module( 2025-09-07T08:14:18.9808985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9809048Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9809274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9809345Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9809570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9809640Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9809897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9810019Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9810022Z 2025-09-07T08:14:18.9810139Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9810321Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9810378Z return mod(**inputs) 2025-09-07T08:14:18.9810606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9810670Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9810899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9810959Z layer_outputs = layer_module( 2025-09-07T08:14:18.9811171Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9811238Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9811465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9811534Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9811759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9811828Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9812052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9812165Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9812169Z 2025-09-07T08:14:18.9812255Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9812438Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9812496Z return mod(**inputs) 2025-09-07T08:14:18.9812723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9812786Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9813016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9813077Z layer_outputs = layer_module( 2025-09-07T08:14:18.9813286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9813390Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9813622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9813689Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9813918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9814021Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9814250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9814357Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9814361Z 2025-09-07T08:14:18.9814449Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9814633Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9814688Z return mod(**inputs) 2025-09-07T08:14:18.9814921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9814981Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9815208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9815319Z layer_outputs = layer_module( 2025-09-07T08:14:18.9815532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9815600Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9815858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9815929Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9816155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9816225Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9816451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9816595Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9816599Z 2025-09-07T08:14:18.9816691Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9816871Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9816926Z return mod(**inputs) 2025-09-07T08:14:18.9817210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9817272Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9817504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9817565Z layer_outputs = layer_module( 2025-09-07T08:14:18.9817776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9817844Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9818068Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9818142Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9818364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9818435Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9818660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9818727Z value_states = self.v(current_states) 2025-09-07T08:14:18.9818731Z 2025-09-07T08:14:18.9818820Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9819041Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9819098Z return mod(**inputs) 2025-09-07T08:14:18.9819325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9819385Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9819653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9819714Z layer_outputs = layer_module( 2025-09-07T08:14:18.9819929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9819993Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9820227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9820296Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9820522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9820592Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9820816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9820949Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9820953Z 2025-09-07T08:14:18.9821043Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9821258Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9821316Z return mod(**inputs) 2025-09-07T08:14:18.9821546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9821610Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9821838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9821901Z layer_outputs = layer_module( 2025-09-07T08:14:18.9822113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9822178Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9822411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9822480Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9822707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9822777Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9823002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9823094Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9823101Z 2025-09-07T08:14:18.9823188Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9823375Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9823429Z return mod(**inputs) 2025-09-07T08:14:18.9823660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9823724Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9823952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9824014Z layer_outputs = layer_module( 2025-09-07T08:14:18.9824225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9824291Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9824528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9824634Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9824870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9824941Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9825213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9825305Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9825308Z 2025-09-07T08:14:18.9825399Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9825587Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9825644Z return mod(**inputs) 2025-09-07T08:14:18.9825881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9825946Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9826175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9826244Z layer_outputs = layer_module( 2025-09-07T08:14:18.9826486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9826562Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9826821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9826893Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9827129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9827200Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9827431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9827521Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9827524Z 2025-09-07T08:14:18.9827622Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9827807Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9827864Z return mod(**inputs) 2025-09-07T08:14:18.9828101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9828168Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9828403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9828463Z layer_outputs = layer_module( 2025-09-07T08:14:18.9828676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9828751Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9828978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9829053Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9829285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9829363Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9829590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9829657Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9829660Z 2025-09-07T08:14:18.9829755Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9829938Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9830034Z return mod(**inputs) 2025-09-07T08:14:18.9830264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9830327Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9830562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9830661Z layer_outputs = layer_module( 2025-09-07T08:14:18.9830879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9830948Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9831171Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9831242Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9831467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 609, in forward 2025-09-07T08:14:18.9831593Z hidden_states = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:14:18.9831596Z 2025-09-07T08:14:18.9831665Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9831754Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9831968Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9832028Z return mod(**inputs) 2025-09-07T08:14:18.9832263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9832359Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9832599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9832663Z layer_outputs = layer_module( 2025-09-07T08:14:18.9832873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9832943Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9833167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9833239Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9833464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T08:14:18.9833558Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9833787Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9833854Z return self.weight * hidden_states 2025-09-07T08:14:18.9833857Z 2025-09-07T08:14:18.9833948Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9834129Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9834187Z return mod(**inputs) 2025-09-07T08:14:18.9834416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9834476Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9834707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9834767Z layer_outputs = layer_module( 2025-09-07T08:14:18.9834977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9835042Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9835267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9835338Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9835568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9835703Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9835930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9835999Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9836059Z 2025-09-07T08:14:18.9836154Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9836343Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9836403Z return mod(**inputs) 2025-09-07T08:14:18.9836633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9836697Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9836927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9836988Z layer_outputs = layer_module( 2025-09-07T08:14:18.9837200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9837271Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9837511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9837624Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9837856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9837967Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9838196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9838277Z key_states = self.k(current_states) 2025-09-07T08:14:18.9838280Z 2025-09-07T08:14:18.9838373Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9838564Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9838621Z return mod(**inputs) 2025-09-07T08:14:18.9838851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9838919Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9839151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9839218Z layer_outputs = layer_module( 2025-09-07T08:14:18.9839433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9839500Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9839728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9839800Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9840028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9840099Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9840324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9840453Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9840457Z 2025-09-07T08:14:18.9840544Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9840736Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9840792Z return mod(**inputs) 2025-09-07T08:14:18.9841028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9841092Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9841358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9841428Z layer_outputs = layer_module( 2025-09-07T08:14:18.9841638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9841738Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9841967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9842038Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9842267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9842338Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9842563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9842673Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9842676Z 2025-09-07T08:14:18.9842765Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9842948Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9843003Z return mod(**inputs) 2025-09-07T08:14:18.9843273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9843337Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9843601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9843665Z layer_outputs = layer_module( 2025-09-07T08:14:18.9843877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9843950Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9844182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9844258Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9844486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9844564Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9844798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9844914Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9844918Z 2025-09-07T08:14:18.9845013Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9845194Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9845251Z return mod(**inputs) 2025-09-07T08:14:18.9845477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9845539Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9845769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9845829Z layer_outputs = layer_module( 2025-09-07T08:14:18.9846047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9846113Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9846339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9846413Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9846639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9846716Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9846997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9847146Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9847152Z 2025-09-07T08:14:18.9847240Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9847459Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9847521Z return mod(**inputs) 2025-09-07T08:14:18.9847753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9847819Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9848048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9848111Z layer_outputs = layer_module( 2025-09-07T08:14:18.9848330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9848396Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9848629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9848731Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9848961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9849036Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9849294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9849369Z value_states = self.v(current_states) 2025-09-07T08:14:18.9849372Z 2025-09-07T08:14:18.9849459Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9849645Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9849703Z return mod(**inputs) 2025-09-07T08:14:18.9849934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9849999Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9850230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9850291Z layer_outputs = layer_module( 2025-09-07T08:14:18.9850505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9850573Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9850809Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9850879Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9851116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9851189Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9851415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9851519Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9851523Z 2025-09-07T08:14:18.9851611Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9851799Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9851857Z return mod(**inputs) 2025-09-07T08:14:18.9852093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9852156Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9852385Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9852492Z layer_outputs = layer_module( 2025-09-07T08:14:18.9852706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9852781Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9853046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9853120Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9853353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9853424Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9853652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9853742Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9853747Z 2025-09-07T08:14:18.9853836Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9854021Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9854079Z return mod(**inputs) 2025-09-07T08:14:18.9854352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9854421Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9854662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9854756Z layer_outputs = layer_module( 2025-09-07T08:14:18.9854971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9855048Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9855274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9855350Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9855576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9855650Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9855881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9855973Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9855976Z 2025-09-07T08:14:18.9856068Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9856248Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9856306Z return mod(**inputs) 2025-09-07T08:14:18.9856543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9856615Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9856847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9856906Z layer_outputs = layer_module( 2025-09-07T08:14:18.9857125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9857246Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9857476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9857555Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9857784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9857866Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9858094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9858267Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9858272Z 2025-09-07T08:14:18.9858359Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9858541Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9858636Z return mod(**inputs) 2025-09-07T08:14:18.9858865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9858933Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9859164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9859230Z layer_outputs = layer_module( 2025-09-07T08:14:18.9859450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9859520Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9859754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9859826Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9860087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9860170Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9860432Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9860509Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9860512Z 2025-09-07T08:14:18.9860586Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9860677Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9860862Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9860922Z return mod(**inputs) 2025-09-07T08:14:18.9861162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9861226Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9861462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9861525Z layer_outputs = layer_module( 2025-09-07T08:14:18.9861739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9861814Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9862039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9862129Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9862356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.9862446Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9862676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9862745Z return self.weight * hidden_states 2025-09-07T08:14:18.9862749Z 2025-09-07T08:14:18.9862849Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9863031Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9863090Z return mod(**inputs) 2025-09-07T08:14:18.9863327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9863389Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9863628Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9863724Z layer_outputs = layer_module( 2025-09-07T08:14:18.9863936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9864001Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9864229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9864341Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9864570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9864678Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9864902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.9864972Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.9864977Z 2025-09-07T08:14:18.9865067Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9865248Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9865307Z return mod(**inputs) 2025-09-07T08:14:18.9865537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9865637Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9865873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9865968Z layer_outputs = layer_module( 2025-09-07T08:14:18.9866188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9866254Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9866483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9866562Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9866786Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9866894Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9867122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.9867195Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.9867199Z 2025-09-07T08:14:18.9867287Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9867468Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9867526Z return mod(**inputs) 2025-09-07T08:14:18.9867753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9867818Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9868048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9868110Z layer_outputs = layer_module( 2025-09-07T08:14:18.9868318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9868386Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9868622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9868701Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9868933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9869031Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9869259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.9869367Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.9869370Z 2025-09-07T08:14:18.9869442Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9869540Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9869724Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9869820Z return mod(**inputs) 2025-09-07T08:14:18.9870059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9870124Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9870361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9870423Z layer_outputs = layer_module( 2025-09-07T08:14:18.9870632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9870700Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9870927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9871007Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9871267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 598, in forward 2025-09-07T08:14:18.9871376Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9871635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9871703Z return self.weight * hidden_states 2025-09-07T08:14:18.9871706Z 2025-09-07T08:14:18.9871797Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9871979Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9872040Z return mod(**inputs) 2025-09-07T08:14:18.9872269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9872331Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9872567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9872630Z layer_outputs = layer_module( 2025-09-07T08:14:18.9872849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9872919Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9873150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9873219Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9873444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9873518Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9873743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9873813Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9873819Z 2025-09-07T08:14:18.9873908Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9874089Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9874152Z return mod(**inputs) 2025-09-07T08:14:18.9874392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9874461Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9874690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9874790Z layer_outputs = layer_module( 2025-09-07T08:14:18.9875019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9875095Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9875335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9875446Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9875680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9875755Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9875979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9876049Z key_states = self.k(current_states) 2025-09-07T08:14:18.9876053Z 2025-09-07T08:14:18.9876148Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9876345Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9876404Z return mod(**inputs) 2025-09-07T08:14:18.9876637Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9876712Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9876975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9877047Z layer_outputs = layer_module( 2025-09-07T08:14:18.9877312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9877384Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9877619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9877691Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9877926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9877996Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9878227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9878353Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9878356Z 2025-09-07T08:14:18.9878451Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9878645Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9878701Z return mod(**inputs) 2025-09-07T08:14:18.9878940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9879006Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9879237Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9879305Z layer_outputs = layer_module( 2025-09-07T08:14:18.9879518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9879592Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9879817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9879889Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9880117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9880188Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9880415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9880559Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9880562Z 2025-09-07T08:14:18.9880657Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9880839Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9880894Z return mod(**inputs) 2025-09-07T08:14:18.9881168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9881233Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9881467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9881529Z layer_outputs = layer_module( 2025-09-07T08:14:18.9881742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9881816Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9882047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9882118Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9882344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9882454Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9882688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9882837Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9882840Z 2025-09-07T08:14:18.9882938Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9883125Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9883190Z return mod(**inputs) 2025-09-07T08:14:18.9883420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9883483Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9883715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9883780Z layer_outputs = layer_module( 2025-09-07T08:14:18.9883998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9884064Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9888422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9888533Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9888804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9888894Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9889139Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9889295Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9889304Z 2025-09-07T08:14:18.9889409Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9889618Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9889682Z return mod(**inputs) 2025-09-07T08:14:18.9889934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9890010Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9890242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9890309Z layer_outputs = layer_module( 2025-09-07T08:14:18.9890600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9890677Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9890919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9890994Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9891264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9891345Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9891572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9891653Z value_states = self.v(current_states) 2025-09-07T08:14:18.9891657Z 2025-09-07T08:14:18.9891753Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9891954Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9892020Z return mod(**inputs) 2025-09-07T08:14:18.9892258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9892328Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9892591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9892660Z layer_outputs = layer_module( 2025-09-07T08:14:18.9892912Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9892985Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9893218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9893289Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9893517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9893590Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9893817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9893927Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9893934Z 2025-09-07T08:14:18.9894031Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9894226Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9894288Z return mod(**inputs) 2025-09-07T08:14:18.9894520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9894590Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9894819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9894891Z layer_outputs = layer_module( 2025-09-07T08:14:18.9895104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9895175Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9895406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9895478Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9895713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9895786Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9896015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9896108Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9896144Z 2025-09-07T08:14:18.9896236Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9896426Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9896482Z return mod(**inputs) 2025-09-07T08:14:18.9896718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9896812Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9897046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9897112Z layer_outputs = layer_module( 2025-09-07T08:14:18.9897408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9897485Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9897724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9897801Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9898032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9898105Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9898377Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9898476Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9898480Z 2025-09-07T08:14:18.9898608Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9898805Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9898868Z return mod(**inputs) 2025-09-07T08:14:18.9899112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9899181Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9899415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9899476Z layer_outputs = layer_module( 2025-09-07T08:14:18.9899692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9899761Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9899988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9900067Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9900295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9900370Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9900597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9900692Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9900696Z 2025-09-07T08:14:18.9900790Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9900976Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9901042Z return mod(**inputs) 2025-09-07T08:14:18.9901275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9901336Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9901573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9901636Z layer_outputs = layer_module( 2025-09-07T08:14:18.9901849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9901951Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9902182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 681, in forward 2025-09-07T08:14:18.9902253Z self_attention_outputs = self.layer[0]( 2025-09-07T08:14:18.9902481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 599, in forward 2025-09-07T08:14:18.9902591Z attention_output = self.SelfAttention( 2025-09-07T08:14:18.9902818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9902894Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9902898Z 2025-09-07T08:14:18.9902973Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9903231Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9903429Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9903490Z return mod(**inputs) 2025-09-07T08:14:18.9903720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9903786Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9904020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9904173Z layer_outputs = layer_module( 2025-09-07T08:14:18.9904392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9904510Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9904744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9904819Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9905044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 634, in forward 2025-09-07T08:14:18.9905142Z normed_hidden_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9905373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9905441Z return self.weight * hidden_states 2025-09-07T08:14:18.9905446Z 2025-09-07T08:14:18.9905543Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9905726Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9905784Z return mod(**inputs) 2025-09-07T08:14:18.9906023Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9906086Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9906318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9906382Z layer_outputs = layer_module( 2025-09-07T08:14:18.9906593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9906660Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9906889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9906966Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9907190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9907268Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9907495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 490, in forward 2025-09-07T08:14:18.9907564Z query_states = self.q(hidden_states) 2025-09-07T08:14:18.9907567Z 2025-09-07T08:14:18.9907660Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9907895Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9907957Z return mod(**inputs) 2025-09-07T08:14:18.9908191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9908299Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9908542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9908604Z layer_outputs = layer_module( 2025-09-07T08:14:18.9908821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9908888Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9909119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9909189Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9909415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9909492Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9909718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 510, in forward 2025-09-07T08:14:18.9909826Z key_states = self.k(current_states) 2025-09-07T08:14:18.9909829Z 2025-09-07T08:14:18.9909917Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9910133Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9910198Z return mod(**inputs) 2025-09-07T08:14:18.9910428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9910492Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9910722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9910781Z layer_outputs = layer_module( 2025-09-07T08:14:18.9910994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9911058Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9911288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9911357Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9911584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9911655Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9911878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9912004Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9912007Z 2025-09-07T08:14:18.9912094Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9912278Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9912332Z return mod(**inputs) 2025-09-07T08:14:18.9912563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9912628Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9912856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9912918Z layer_outputs = layer_module( 2025-09-07T08:14:18.9913127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9913192Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9913466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9913540Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9913767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9913887Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9914117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9914230Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9914233Z 2025-09-07T08:14:18.9914321Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9914506Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9914560Z return mod(**inputs) 2025-09-07T08:14:18.9914789Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9914852Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9915079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9915141Z layer_outputs = layer_module( 2025-09-07T08:14:18.9915384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9915454Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9915711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9915785Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9916012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9916086Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9916317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 526, in forward 2025-09-07T08:14:18.9916425Z scores = torch.matmul(query_states, key_states.transpose(3, 2)) 2025-09-07T08:14:18.9916428Z 2025-09-07T08:14:18.9916515Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9916698Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9916753Z return mod(**inputs) 2025-09-07T08:14:18.9916986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9917048Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9917282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9917341Z layer_outputs = layer_module( 2025-09-07T08:14:18.9917552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9917626Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9917853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9917924Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9918151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9918223Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9918447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 558, in forward 2025-09-07T08:14:18.9918595Z attn_weights = nn.functional.softmax(scores.float(), dim=-1).type_as(scores) 2025-09-07T08:14:18.9918599Z 2025-09-07T08:14:18.9918690Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9918872Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9918967Z return mod(**inputs) 2025-09-07T08:14:18.9919205Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9919268Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9919542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9919607Z layer_outputs = layer_module( 2025-09-07T08:14:18.9919821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9919887Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9920114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9920184Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9920409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9920481Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9920706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 511, in forward 2025-09-07T08:14:18.9920810Z value_states = self.v(current_states) 2025-09-07T08:14:18.9920814Z 2025-09-07T08:14:18.9920902Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9921084Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9921174Z return mod(**inputs) 2025-09-07T08:14:18.9921406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9921470Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9921701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9921765Z layer_outputs = layer_module( 2025-09-07T08:14:18.9921978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9922043Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9922275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9922343Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9922571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9922642Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9922866Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9922967Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9922972Z 2025-09-07T08:14:18.9923061Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9923250Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9923307Z return mod(**inputs) 2025-09-07T08:14:18.9923541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9923612Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9923841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9923911Z layer_outputs = layer_module( 2025-09-07T08:14:18.9924124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9924189Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9924425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9924528Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9924762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9924834Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9925122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9925222Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9925226Z 2025-09-07T08:14:18.9925318Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9925514Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9925573Z return mod(**inputs) 2025-09-07T08:14:18.9925816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9925884Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9926117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9926185Z layer_outputs = layer_module( 2025-09-07T08:14:18.9926434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9926513Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9926742Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9926848Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9927084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9927157Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9927384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 565, in forward 2025-09-07T08:14:18.9927476Z attn_output = torch.matmul(attn_weights, value_states) 2025-09-07T08:14:18.9927480Z 2025-09-07T08:14:18.9927570Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9927756Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9927816Z return mod(**inputs) 2025-09-07T08:14:18.9928047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9928110Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9928339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9928399Z layer_outputs = layer_module( 2025-09-07T08:14:18.9928610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9928677Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9928902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9928973Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9929197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9929272Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9929499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 567, in forward 2025-09-07T08:14:18.9929591Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:14:18.9929594Z 2025-09-07T08:14:18.9929683Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9929867Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9929961Z return mod(**inputs) 2025-09-07T08:14:18.9930191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9930252Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9930482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9930579Z layer_outputs = layer_module( 2025-09-07T08:14:18.9930795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9930862Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9931091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9931164Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9931391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 635, in forward 2025-09-07T08:14:18.9931467Z attention_output = self.EncDecAttention( 2025-09-07T08:14:18.9931692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 569, in forward 2025-09-07T08:14:18.9931764Z attn_output = self.o(attn_output) 2025-09-07T08:14:18.9931767Z 2025-09-07T08:14:18.9931888Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9932070Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9932131Z return mod(**inputs) 2025-09-07T08:14:18.9932394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9932460Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9932687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9932747Z layer_outputs = layer_module( 2025-09-07T08:14:18.9932964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9933029Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9933259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 705, in forward 2025-09-07T08:14:18.9933332Z cross_attention_outputs = self.layer[1]( 2025-09-07T08:14:18.9933557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 647, in forward 2025-09-07T08:14:18.9933684Z layer_output = hidden_states + self.dropout(attention_output[0]) 2025-09-07T08:14:18.9933687Z 2025-09-07T08:14:18.9933760Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9933851Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9934031Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9934093Z return mod(**inputs) 2025-09-07T08:14:18.9934323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9934384Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9934619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9934686Z layer_outputs = layer_module( 2025-09-07T08:14:18.9934900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9934967Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9935192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9935278Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9935505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 341, in forward 2025-09-07T08:14:18.9935626Z forwarded_states = self.layer_norm(hidden_states) 2025-09-07T08:14:18.9935852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 261, in forward 2025-09-07T08:14:18.9935928Z return self.weight * hidden_states 2025-09-07T08:14:18.9935931Z 2025-09-07T08:14:18.9936062Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9936243Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9936308Z return mod(**inputs) 2025-09-07T08:14:18.9936541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9936611Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9936840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9936901Z layer_outputs = layer_module( 2025-09-07T08:14:18.9937123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9937235Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9937466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9937581Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9937820Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9937970Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9938204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 287, in forward 2025-09-07T08:14:18.9938287Z hidden_states = self.wi(hidden_states) 2025-09-07T08:14:18.9938290Z 2025-09-07T08:14:18.9938388Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9938585Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9938643Z return mod(**inputs) 2025-09-07T08:14:18.9938878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9938955Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9939186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9939258Z layer_outputs = layer_module( 2025-09-07T08:14:18.9939477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9939548Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9939785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9939866Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9940097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9940203Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9940432Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 288, in forward 2025-09-07T08:14:18.9940514Z hidden_states = self.act(hidden_states) 2025-09-07T08:14:18.9940517Z 2025-09-07T08:14:18.9940608Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9940802Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9940857Z return mod(**inputs) 2025-09-07T08:14:18.9941094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1762, in forward 2025-09-07T08:14:18.9941157Z decoder_outputs = self.decoder( 2025-09-07T08:14:18.9941437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1092, in forward 2025-09-07T08:14:18.9941499Z layer_outputs = layer_module( 2025-09-07T08:14:18.9941712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:14:18.9941816Z return super().__call__(*args, **kwargs) 2025-09-07T08:14:18.9942046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 731, in forward 2025-09-07T08:14:18.9942123Z hidden_states = self.layer[-1](hidden_states) 2025-09-07T08:14:18.9942347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 342, in forward 2025-09-07T08:14:18.9942446Z forwarded_states = self.DenseReluDense(forwarded_states) 2025-09-07T08:14:18.9942675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 296, in forward 2025-09-07T08:14:18.9942746Z hidden_states = self.wo(hidden_states) 2025-09-07T08:14:18.9942749Z 2025-09-07T08:14:18.9942822Z cudagraph partition due to non gpu ops 2025-09-07T08:14:18.9942910Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9943093Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9943184Z return mod(**inputs) 2025-09-07T08:14:18.9943415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1789, in forward 2025-09-07T08:14:18.9943558Z sequence_output = sequence_output * (self.model_dim**-0.5) 2025-09-07T08:14:18.9943561Z 2025-09-07T08:14:18.9943649Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:14:18.9943835Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:14:18.9943890Z return mod(**inputs) 2025-09-07T08:14:18.9944119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/t5/modeling_t5.py", line 1791, in forward 2025-09-07T08:14:18.9944197Z lm_logits = self.lm_head(sequence_output) 2025-09-07T08:14:18.9944200Z 2025-09-07T08:15:17.2550936Z pass 2025-09-07T08:15:17.2551349Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:15:24.2042603Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:15:24.2043487Z import pynvml # type: ignore[import] 2025-09-07T08:15:26.2187614Z 2025-09-07T08:15:38.5646677Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:15:38.5647018Z loading model: 0it [00:12, ?it/s] 2025-09-07T08:15:38.6514908Z cpu eval hf_T5_large 2025-09-07T08:15:39.0497228Z pass_due_to_skip 2025-09-07T08:15:39.0500196Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:15:41.3696871Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:15:41.3697826Z import pynvml # type: ignore[import] 2025-09-07T08:15:43.3787852Z 2025-09-07T08:15:49.6300198Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:15:49.6300541Z loading model: 0it [00:06, ?it/s] 2025-09-07T08:15:49.6687429Z cpu eval hf_distil_whisper 2025-09-07T08:15:56.9696140Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:15:59.0598346Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:16:01.1537164Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:16:09.7600354Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7600822Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7601192Z return mod(*inputs) 2025-09-07T08:16:09.7602033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7602471Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7602895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 677, in forward 2025-09-07T08:16:09.7603488Z inputs_embeds = nn.functional.gelu(self.conv1(input_features)) 2025-09-07T08:16:09.7603679Z 2025-09-07T08:16:09.7603797Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7604168Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7604493Z return mod(*inputs) 2025-09-07T08:16:09.7604879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7605285Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7605799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 677, in forward 2025-09-07T08:16:09.7606257Z inputs_embeds = nn.functional.gelu(self.conv1(input_features)) 2025-09-07T08:16:09.7606432Z 2025-09-07T08:16:09.7606617Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7606982Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7607309Z return mod(*inputs) 2025-09-07T08:16:09.7607696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7608106Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7608501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 678, in forward 2025-09-07T08:16:09.7608950Z inputs_embeds = nn.functional.gelu(self.conv2(inputs_embeds)) 2025-09-07T08:16:09.7609136Z 2025-09-07T08:16:09.7609245Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7609600Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7609921Z return mod(*inputs) 2025-09-07T08:16:09.7610278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7610676Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7611060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7611458Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7611809Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7612190Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7612595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 409, in forward 2025-09-07T08:16:09.7613034Z hidden_states = self.self_attn_layer_norm(hidden_states) 2025-09-07T08:16:09.7613199Z 2025-09-07T08:16:09.7613396Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7613758Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7614096Z return mod(*inputs) 2025-09-07T08:16:09.7614466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7615787Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7616182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7616580Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7616925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7617349Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7617821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 409, in forward 2025-09-07T08:16:09.7618266Z hidden_states = self.self_attn_layer_norm(hidden_states) 2025-09-07T08:16:09.7618433Z 2025-09-07T08:16:09.7618536Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7618891Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7619217Z return mod(*inputs) 2025-09-07T08:16:09.7619598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7620010Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7620395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7620838Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7621188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7621556Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7621982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7622405Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7622830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.7623269Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.7623431Z 2025-09-07T08:16:09.7623537Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7623883Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7624210Z return mod(*inputs) 2025-09-07T08:16:09.7624579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7624981Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7625367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7625756Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7626098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7626458Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7626856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7627273Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7627686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.7628180Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.7628402Z 2025-09-07T08:16:09.7628499Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7628873Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7629188Z return mod(*inputs) 2025-09-07T08:16:09.7629546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7629991Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7630375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7630771Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7631101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7631503Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7631894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7632306Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7632708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.7633191Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.7633423Z 2025-09-07T08:16:09.7633499Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7633706Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7633906Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7634092Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7634312Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7634694Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7635026Z return mod(*inputs) 2025-09-07T08:16:09.7635438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7635844Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7636234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7636628Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7636976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7637329Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7637709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7638122Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7638529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.7638946Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.7639389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.7639877Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.7640065Z 2025-09-07T08:16:09.7640159Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7640516Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7640833Z return mod(*inputs) 2025-09-07T08:16:09.7641188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7641608Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7641989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7642381Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7642722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7643068Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7643470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7643925Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7644339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.7644760Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.7645200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.7645706Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.7645878Z 2025-09-07T08:16:09.7645979Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7646337Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7646650Z return mod(*inputs) 2025-09-07T08:16:09.7647037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7647439Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7647834Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7648237Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7648664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7649041Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7649521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7649971Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7650407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.7650808Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.7650951Z 2025-09-07T08:16:09.7651063Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7651425Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7651753Z return mod(*inputs) 2025-09-07T08:16:09.7652132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7652539Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7652949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7653349Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7653723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7654088Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7654493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 420, in forward 2025-09-07T08:16:09.7654928Z hidden_states = self.final_layer_norm(hidden_states) 2025-09-07T08:16:09.7655092Z 2025-09-07T08:16:09.7655205Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7655571Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7655889Z return mod(*inputs) 2025-09-07T08:16:09.7656264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7656665Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7657052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7657451Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7657857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7658292Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7658695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.7659137Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.7659305Z 2025-09-07T08:16:09.7659452Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7659805Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7660126Z return mod(*inputs) 2025-09-07T08:16:09.7660496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7660893Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7661273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7661666Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7662002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7662364Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7662755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.7663215Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.7663598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.7663971Z return self.act(input) 2025-09-07T08:16:09.7664080Z 2025-09-07T08:16:09.7664185Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7664529Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7664842Z return mod(*inputs) 2025-09-07T08:16:09.7665217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7665619Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7665998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7666392Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7666736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7667089Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7667479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.7667893Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.7668028Z 2025-09-07T08:16:09.7668127Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7668476Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7668802Z return mod(*inputs) 2025-09-07T08:16:09.7669164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7669567Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7669950Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7670342Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7670685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7671052Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7671441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 409, in forward 2025-09-07T08:16:09.7671874Z hidden_states = self.self_attn_layer_norm(hidden_states) 2025-09-07T08:16:09.7672088Z 2025-09-07T08:16:09.7672182Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7672524Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7672843Z return mod(*inputs) 2025-09-07T08:16:09.7673201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7673637Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7674022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7674415Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7674754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7675103Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7675496Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7675913Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7676332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.7676802Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.7676969Z 2025-09-07T08:16:09.7677063Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7677460Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7677785Z return mod(*inputs) 2025-09-07T08:16:09.7678169Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7678562Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7678959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7679360Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7679714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7680085Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7680488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7680902Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7681317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.7681807Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.7682024Z 2025-09-07T08:16:09.7682130Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7682482Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7682799Z return mod(*inputs) 2025-09-07T08:16:09.7683170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7683573Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7683954Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7684350Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7684697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7685065Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7685457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7685862Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7686308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.7686793Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.7687013Z 2025-09-07T08:16:09.7687133Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7687339Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7687526Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7687721Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7687945Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7688298Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7688608Z return mod(*inputs) 2025-09-07T08:16:09.7688971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7689373Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7689753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7690145Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7690511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7690874Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7691301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7691713Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7692109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.7692528Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.7692980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.7693463Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.7693643Z 2025-09-07T08:16:09.7693748Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7694088Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7694404Z return mod(*inputs) 2025-09-07T08:16:09.7694767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7695161Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7695538Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7695921Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7696260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7696620Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7697010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7697406Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7697864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.7698278Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.7698743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.7699203Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.7699362Z 2025-09-07T08:16:09.7699455Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7699844Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7700155Z return mod(*inputs) 2025-09-07T08:16:09.7700515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7700915Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7701327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7701719Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7702054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7702406Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7702793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7703356Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7703769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.7704173Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.7704302Z 2025-09-07T08:16:09.7704401Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7704826Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7705150Z return mod(*inputs) 2025-09-07T08:16:09.7705582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7706007Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7706407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7706797Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7707143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7707503Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7707896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:16:09.7708304Z hidden_states = residual + hidden_states 2025-09-07T08:16:09.7708432Z 2025-09-07T08:16:09.7708531Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7708882Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7709197Z return mod(*inputs) 2025-09-07T08:16:09.7709557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7709947Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7710331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7710725Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7711064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7711412Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7711795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.7712235Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.7712407Z 2025-09-07T08:16:09.7712503Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7712843Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7713154Z return mod(*inputs) 2025-09-07T08:16:09.7713509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7713963Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7714345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7714731Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7715114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7715469Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7715860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.7716302Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.7716684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.7717015Z return self.act(input) 2025-09-07T08:16:09.7717128Z 2025-09-07T08:16:09.7717223Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7717564Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7717878Z return mod(*inputs) 2025-09-07T08:16:09.7718270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7718675Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7719059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7719483Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7719821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7720170Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7720563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.7720965Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.7721096Z 2025-09-07T08:16:09.7721195Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7721532Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7721843Z return mod(*inputs) 2025-09-07T08:16:09.7722199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7722597Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7722977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7723364Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7723702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7724053Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7724441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7724848Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7725248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.7725684Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.7725856Z 2025-09-07T08:16:09.7725949Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7726289Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7726601Z return mod(*inputs) 2025-09-07T08:16:09.7726951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7727392Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7727772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7728164Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7728497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7728890Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7729286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7729698Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7730101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.7730581Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.7730827Z 2025-09-07T08:16:09.7730923Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7731261Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7731579Z return mod(*inputs) 2025-09-07T08:16:09.7743485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7743994Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7744477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7744940Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7745306Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7745680Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7746094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7746525Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7746946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.7747446Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.7747689Z 2025-09-07T08:16:09.7747771Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7747978Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7748176Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7748368Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7748596Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7748963Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7749297Z return mod(*inputs) 2025-09-07T08:16:09.7749699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7750107Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7750527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7750939Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7751291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7751660Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7752066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7752490Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7752907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.7753379Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.7753829Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.7754338Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.7754565Z 2025-09-07T08:16:09.7754671Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7755036Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7755363Z return mod(*inputs) 2025-09-07T08:16:09.7755740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7756154Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7756542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7756937Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7757283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7757640Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7758070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7758491Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7758932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.7759357Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.7759794Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.7760259Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.7760427Z 2025-09-07T08:16:09.7760525Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7760879Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7761190Z return mod(*inputs) 2025-09-07T08:16:09.7761560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7761960Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7762348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7762743Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7763082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7763436Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7763829Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7764240Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7764641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.7765037Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.7765176Z 2025-09-07T08:16:09.7765271Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7765618Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7765936Z return mod(*inputs) 2025-09-07T08:16:09.7766294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7766692Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7767073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7767514Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7767853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7768203Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7768631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.7769072Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.7769241Z 2025-09-07T08:16:09.7769344Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7769688Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7769996Z return mod(*inputs) 2025-09-07T08:16:09.7770355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7770752Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7771131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7771520Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7771888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7772253Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7772682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.7773123Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.7773500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.7773841Z return self.act(input) 2025-09-07T08:16:09.7773955Z 2025-09-07T08:16:09.7774053Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7774395Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7774708Z return mod(*inputs) 2025-09-07T08:16:09.7775063Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7775461Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7775850Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7776250Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7776577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7776928Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7777319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.7777793Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.7777928Z 2025-09-07T08:16:09.7778028Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7778368Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7778683Z return mod(*inputs) 2025-09-07T08:16:09.7779040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7779438Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7779815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7780203Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7780536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7780953Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7781347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7781755Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7782165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.7782636Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.7782799Z 2025-09-07T08:16:09.7782899Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7783246Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7783554Z return mod(*inputs) 2025-09-07T08:16:09.7783914Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7784307Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7784690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7785078Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7785402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7785788Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7786179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7786641Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7787045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.7787529Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.7787756Z 2025-09-07T08:16:09.7787851Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7788201Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7788514Z return mod(*inputs) 2025-09-07T08:16:09.7788867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7789267Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7789652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7790046Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7790381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7790732Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7791119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7791526Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7791933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.7792420Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.7792644Z 2025-09-07T08:16:09.7792721Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7792926Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7793124Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7793319Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7793530Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7793870Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7794182Z return mod(*inputs) 2025-09-07T08:16:09.7794543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7794999Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7795382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7795775Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7796156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7796503Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7796884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7797293Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7797685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.7798111Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.7798548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.7799020Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.7799203Z 2025-09-07T08:16:09.7799335Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7799678Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7799978Z return mod(*inputs) 2025-09-07T08:16:09.7800956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7801382Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7801785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7802182Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7802510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7802858Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7803444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7803875Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7804297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.7804722Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.7805174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.7805631Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.7805789Z 2025-09-07T08:16:09.7805889Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7806245Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7806560Z return mod(*inputs) 2025-09-07T08:16:09.7806932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7807327Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7807701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7808085Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7808427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7808772Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7809159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7809650Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7810043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.7810441Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.7810621Z 2025-09-07T08:16:09.7810716Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7811056Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7811379Z return mod(*inputs) 2025-09-07T08:16:09.7811738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7812130Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7812508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7812898Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7813229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7813588Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7814041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:16:09.7814468Z hidden_states = residual + hidden_states 2025-09-07T08:16:09.7814601Z 2025-09-07T08:16:09.7814706Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7815100Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7815423Z return mod(*inputs) 2025-09-07T08:16:09.7815790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7816192Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7816571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7816961Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7817299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7817696Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7818084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.7818519Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.7818687Z 2025-09-07T08:16:09.7818781Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7819118Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7819425Z return mod(*inputs) 2025-09-07T08:16:09.7819781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7820165Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7820540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7820929Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7821269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7821619Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7822013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.7822448Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.7822826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.7823206Z return self.act(input) 2025-09-07T08:16:09.7823310Z 2025-09-07T08:16:09.7823404Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7823745Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7824055Z return mod(*inputs) 2025-09-07T08:16:09.7824454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7824849Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7825233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7825619Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7825955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7826303Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7826687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.7827080Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.7827213Z 2025-09-07T08:16:09.7827310Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7827684Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7827999Z return mod(*inputs) 2025-09-07T08:16:09.7828381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7828775Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7829151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7829535Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7829863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7830209Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7830609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7831020Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7831427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.7831852Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.7832015Z 2025-09-07T08:16:09.7832113Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7832455Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7832767Z return mod(*inputs) 2025-09-07T08:16:09.7833125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7833517Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7833895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7834279Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7834616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7834964Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7835347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7835749Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7836146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.7836666Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.7836881Z 2025-09-07T08:16:09.7837202Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7837541Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7837849Z return mod(*inputs) 2025-09-07T08:16:09.7838244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7838632Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7839007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7839390Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7839813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7840162Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7840550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7840950Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7841454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.7841938Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.7842159Z 2025-09-07T08:16:09.7842236Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7842489Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7842677Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7842864Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7843075Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7843414Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7843724Z return mod(*inputs) 2025-09-07T08:16:09.7844079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7844469Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7844845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7845232Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7845566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7845917Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7846306Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7846702Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7847095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.7847510Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.7847954Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.7848435Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.7848614Z 2025-09-07T08:16:09.7848712Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7849053Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7849366Z return mod(*inputs) 2025-09-07T08:16:09.7849718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7850113Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7850482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7850924Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7851258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7851609Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7852039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7852441Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7852848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.7853258Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.7853696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.7854152Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.7854309Z 2025-09-07T08:16:09.7854404Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7854745Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7855059Z return mod(*inputs) 2025-09-07T08:16:09.7855455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7855856Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7856265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7856655Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7856986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7857336Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7857762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7858168Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7858567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.7858965Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.7859090Z 2025-09-07T08:16:09.7859187Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7859520Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7859828Z return mod(*inputs) 2025-09-07T08:16:09.7860181Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7860573Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7860949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7861329Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7861656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7862004Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7862382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.7862812Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.7862980Z 2025-09-07T08:16:09.7863071Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7863411Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7863717Z return mod(*inputs) 2025-09-07T08:16:09.7864123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7864506Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7864879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7865305Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7865638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7865983Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7866365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.7866817Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.7867189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.7867532Z return self.act(input) 2025-09-07T08:16:09.7867634Z 2025-09-07T08:16:09.7867729Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7868059Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7868366Z return mod(*inputs) 2025-09-07T08:16:09.7868755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7869145Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7869551Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7869940Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7870274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7870630Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7871011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.7871402Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.7871534Z 2025-09-07T08:16:09.7871626Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7871966Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7872277Z return mod(*inputs) 2025-09-07T08:16:09.7872623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7873022Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7873393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7873783Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7874106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7874451Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7874843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7875267Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7875669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.7876090Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.7876254Z 2025-09-07T08:16:09.7876345Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7876688Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7876992Z return mod(*inputs) 2025-09-07T08:16:09.7877338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7877776Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7878145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7878558Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7878998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7879341Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7879719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7880139Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7880532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.7881022Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.7881239Z 2025-09-07T08:16:09.7881332Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7881673Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7881983Z return mod(*inputs) 2025-09-07T08:16:09.7882387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7882788Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7883207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7883600Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7883952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7884312Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7884720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7885123Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7885522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.7886015Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.7886235Z 2025-09-07T08:16:09.7886313Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7886512Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7886710Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7886902Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7887123Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7887467Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7887799Z return mod(*inputs) 2025-09-07T08:16:09.7888159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7888564Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7888951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7889339Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7889667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7890019Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7890427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7890829Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7891223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.7891681Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.7892124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.7892652Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.7892834Z 2025-09-07T08:16:09.7892930Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7893285Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7893605Z return mod(*inputs) 2025-09-07T08:16:09.7893964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7894362Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7894743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7895126Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7895455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7895800Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7896246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7896670Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7897095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.7897513Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.7897985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.7898450Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.7898607Z 2025-09-07T08:16:09.7898701Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7899038Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7899348Z return mod(*inputs) 2025-09-07T08:16:09.7899705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7900100Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7900470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7900860Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7901193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7901547Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7901937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7902332Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7902731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.7903270Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.7903395Z 2025-09-07T08:16:09.7903487Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7903823Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7904137Z return mod(*inputs) 2025-09-07T08:16:09.7904489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7904889Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7905367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7905750Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7906077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7906489Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7906879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:16:09.7907283Z hidden_states = residual + hidden_states 2025-09-07T08:16:09.7907411Z 2025-09-07T08:16:09.7907504Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7907852Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7908168Z return mod(*inputs) 2025-09-07T08:16:09.7908526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7908918Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7909301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7909686Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7910079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7910430Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7910851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.7911290Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.7911461Z 2025-09-07T08:16:09.7911554Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7911890Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7912205Z return mod(*inputs) 2025-09-07T08:16:09.7912556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7912951Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7913323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7913709Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7914034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7914386Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7914767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.7915198Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.7915585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.7915921Z return self.act(input) 2025-09-07T08:16:09.7916033Z 2025-09-07T08:16:09.7916125Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7916473Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7916787Z return mod(*inputs) 2025-09-07T08:16:09.7917160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7917550Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7917924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7918313Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7918660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7919058Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7919455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.7919855Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.7920027Z 2025-09-07T08:16:09.7920135Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7920480Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7920784Z return mod(*inputs) 2025-09-07T08:16:09.7921149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7921544Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7921926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7922316Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7922641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7923006Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7923429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7923853Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7924283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.7924720Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.7924882Z 2025-09-07T08:16:09.7924972Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7925307Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7925623Z return mod(*inputs) 2025-09-07T08:16:09.7925978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7926380Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7926771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7927166Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7927502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7927849Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7928233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7928636Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7929042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.7929518Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.7929736Z 2025-09-07T08:16:09.7929834Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7930192Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7930524Z return mod(*inputs) 2025-09-07T08:16:09.7930895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7931294Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7931686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7932082Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7932427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7932833Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7933228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7933651Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7934099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.7934588Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.7934814Z 2025-09-07T08:16:09.7934902Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7935103Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7935300Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7935495Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7935718Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7936067Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7936387Z return mod(*inputs) 2025-09-07T08:16:09.7936753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7937157Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7937648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7938048Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7938426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7938796Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7939192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7939601Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7940010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.7940431Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.7940889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.7941374Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.7941556Z 2025-09-07T08:16:09.7941654Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7942004Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7942324Z return mod(*inputs) 2025-09-07T08:16:09.7942689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7943090Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7943462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7943857Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7944204Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7944562Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7944949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7945360Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7945762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.7946179Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.7946622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.7947117Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.7947282Z 2025-09-07T08:16:09.7947378Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7947729Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7948101Z return mod(*inputs) 2025-09-07T08:16:09.7948464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7948857Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7949240Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7949644Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7950016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7950381Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7950789Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7951202Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7951654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.7952056Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.7952184Z 2025-09-07T08:16:09.7952314Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7952669Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7952992Z return mod(*inputs) 2025-09-07T08:16:09.7953360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7953770Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7954148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7954541Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7954877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7955236Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7955618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.7956052Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.7956219Z 2025-09-07T08:16:09.7956312Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7956649Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7956956Z return mod(*inputs) 2025-09-07T08:16:09.7957308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7957700Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7958077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7958463Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7958792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7959133Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7959513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.7959946Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.7960330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.7960749Z return self.act(input) 2025-09-07T08:16:09.7960857Z 2025-09-07T08:16:09.7960951Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7961296Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7961658Z return mod(*inputs) 2025-09-07T08:16:09.7962021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7962411Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7962795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7963186Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7963532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7963888Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7964275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.7964685Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.7964826Z 2025-09-07T08:16:09.7964956Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7965303Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7965611Z return mod(*inputs) 2025-09-07T08:16:09.7966003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7966410Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7966797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7967195Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7967526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7967885Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7968273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7968689Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7969097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.7969527Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.7969696Z 2025-09-07T08:16:09.7969789Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7970132Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7970446Z return mod(*inputs) 2025-09-07T08:16:09.7970799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7971194Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7971574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7971971Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7972310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7972657Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7973047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7973455Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7973860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.7974393Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.7974612Z 2025-09-07T08:16:09.7974706Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7975049Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7975403Z return mod(*inputs) 2025-09-07T08:16:09.7975763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7976165Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7976540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7976936Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7977275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7977663Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7978051Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7978460Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7978915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.7979414Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.7979637Z 2025-09-07T08:16:09.7979754Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7979954Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7980149Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7980343Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.7980560Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7980900Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7981212Z return mod(*inputs) 2025-09-07T08:16:09.7981573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7981968Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7982355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7982743Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7983079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7983433Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7983823Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7984229Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7984627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.7985044Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.7985498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.7985981Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.7986162Z 2025-09-07T08:16:09.7986258Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7986613Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7986928Z return mod(*inputs) 2025-09-07T08:16:09.7987294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7987692Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7988120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7988520Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7988857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7989252Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7989640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7990040Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7990445Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.7990858Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.7991299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.7991763Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.7991925Z 2025-09-07T08:16:09.7992019Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7992364Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7992727Z return mod(*inputs) 2025-09-07T08:16:09.7993087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7993511Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7993893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7994282Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7994501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7994585Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7994847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.7994932Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.7995192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.7995268Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.7995276Z 2025-09-07T08:16:09.7995371Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7995560Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7995628Z return mod(*inputs) 2025-09-07T08:16:09.7995894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7995967Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7996228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7996294Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7996518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7996588Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7996858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:16:09.7996930Z hidden_states = residual + hidden_states 2025-09-07T08:16:09.7996933Z 2025-09-07T08:16:09.7997025Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7997216Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7997272Z return mod(*inputs) 2025-09-07T08:16:09.7997611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7997680Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7997942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7998045Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.7998264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.7998345Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.7998602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.7998716Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.7998719Z 2025-09-07T08:16:09.7998810Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.7998998Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.7999065Z return mod(*inputs) 2025-09-07T08:16:09.7999329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.7999435Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.7999699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.7999772Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8000022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8000096Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8000365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8000475Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8000694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.8000761Z return self.act(input) 2025-09-07T08:16:09.8000764Z 2025-09-07T08:16:09.8000857Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8001056Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8001116Z return mod(*inputs) 2025-09-07T08:16:09.8001388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8001453Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8001712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8001784Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8002003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8002079Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8002341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.8002427Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.8002430Z 2025-09-07T08:16:09.8002523Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8002710Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8002776Z return mod(*inputs) 2025-09-07T08:16:09.8003172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8003247Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8003518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8003681Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8003906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8003977Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8004300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8004389Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8004662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.8004768Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.8004772Z 2025-09-07T08:16:09.8004868Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8005068Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8005130Z return mod(*inputs) 2025-09-07T08:16:09.8005400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8005466Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8005795Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8005870Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8006141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8006223Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8006484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8006568Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8006840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.8006996Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8007000Z 2025-09-07T08:16:09.8007099Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8007289Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8007354Z return mod(*inputs) 2025-09-07T08:16:09.8007624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8007691Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8007958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8008024Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8008250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8008321Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8008579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8008668Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8008926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.8009095Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8009098Z 2025-09-07T08:16:09.8009173Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8009250Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8009321Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8009392Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8009540Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8009728Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8009794Z return mod(*inputs) 2025-09-07T08:16:09.8010064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8010166Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8010434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8010502Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8018714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8018838Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8019154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8019254Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8019549Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8019653Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8020050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.8020188Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.8020229Z 2025-09-07T08:16:09.8020339Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8020556Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8020621Z return mod(*inputs) 2025-09-07T08:16:09.8020913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8020993Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8021277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8021348Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8021575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8021655Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8021926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8022017Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8022287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8022385Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8022687Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.8022790Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.8022794Z 2025-09-07T08:16:09.8022899Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8023107Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8023168Z return mod(*inputs) 2025-09-07T08:16:09.8023447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8023518Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8023781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8023847Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8024111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8024191Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8024454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8024583Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8024844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.8024921Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.8024925Z 2025-09-07T08:16:09.8025025Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8025219Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8025279Z return mod(*inputs) 2025-09-07T08:16:09.8025550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8025617Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8025877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8025977Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8026200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8026269Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8026580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8026699Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8026703Z 2025-09-07T08:16:09.8026796Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8026995Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8027054Z return mod(*inputs) 2025-09-07T08:16:09.8027325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8027389Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8027652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8027722Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8027946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8028019Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8028279Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8028387Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8028602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.8028668Z return self.act(input) 2025-09-07T08:16:09.8028672Z 2025-09-07T08:16:09.8028771Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8028967Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8029032Z return mod(*inputs) 2025-09-07T08:16:09.8029311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8029377Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8029647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8029712Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8029974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8030045Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8030314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.8030426Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.8030431Z 2025-09-07T08:16:09.8030528Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8030724Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8030785Z return mod(*inputs) 2025-09-07T08:16:09.8031060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8031125Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8031385Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8031458Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8031673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8031747Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8032045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8032134Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8032475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.8032584Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.8032588Z 2025-09-07T08:16:09.8032688Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8032875Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8032942Z return mod(*inputs) 2025-09-07T08:16:09.8033208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8033274Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8033541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8033609Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8033830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8033898Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8034158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8034245Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8034506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.8034671Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8034674Z 2025-09-07T08:16:09.8034766Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8034965Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8035024Z return mod(*inputs) 2025-09-07T08:16:09.8035290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8035361Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8035621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8035690Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8035945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8036014Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8036281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8036394Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8036658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.8036820Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8036824Z 2025-09-07T08:16:09.8036901Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8036970Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8037038Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8037111Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8037202Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8037393Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8037450Z return mod(*inputs) 2025-09-07T08:16:09.8037710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8037810Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8038070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8038170Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8038385Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8038456Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8038715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8038797Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8039063Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8039155Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8039456Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.8039582Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.8039586Z 2025-09-07T08:16:09.8039676Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8039866Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8039922Z return mod(*inputs) 2025-09-07T08:16:09.8040189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8040254Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8040513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8040578Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8040793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8040866Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8041123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8041202Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8041462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8041544Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8041874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.8041973Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.8041976Z 2025-09-07T08:16:09.8042069Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8042288Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8042347Z return mod(*inputs) 2025-09-07T08:16:09.8042614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8042677Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8042941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8043003Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8043222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8043292Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8043562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8043680Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8043943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.8044052Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.8044056Z 2025-09-07T08:16:09.8044148Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8044334Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8044395Z return mod(*inputs) 2025-09-07T08:16:09.8044660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8044734Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8044994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8045060Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8045277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8045347Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8045609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:16:09.8045679Z hidden_states = residual + hidden_states 2025-09-07T08:16:09.8045683Z 2025-09-07T08:16:09.8045778Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8045962Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8046021Z return mod(*inputs) 2025-09-07T08:16:09.8046285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8046351Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8046617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8046682Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8046899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8046972Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8047236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8047349Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8047415Z 2025-09-07T08:16:09.8047504Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8047689Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8047745Z return mod(*inputs) 2025-09-07T08:16:09.8048007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8048109Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8048371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8048439Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8048649Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8048717Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8048976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8049080Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8049294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.8049356Z return self.act(input) 2025-09-07T08:16:09.8049361Z 2025-09-07T08:16:09.8049505Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8049692Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8049782Z return mod(*inputs) 2025-09-07T08:16:09.8050048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8050111Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8050376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8050440Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8050655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8050723Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8050981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.8051057Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.8051060Z 2025-09-07T08:16:09.8051149Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8051332Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8051394Z return mod(*inputs) 2025-09-07T08:16:09.8051654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8051721Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8051977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8052041Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8052250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8052319Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8052578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8052658Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8052916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.8053017Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.8053020Z 2025-09-07T08:16:09.8053163Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8053351Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8053407Z return mod(*inputs) 2025-09-07T08:16:09.8053670Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8053775Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8054036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8054100Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8054312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8054385Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8054645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8054725Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8054982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.8055133Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8055141Z 2025-09-07T08:16:09.8055271Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8055456Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8055546Z return mod(*inputs) 2025-09-07T08:16:09.8058293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8058384Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8058667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8058741Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8058967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8059042Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8059315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8059414Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8059683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.8059858Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8059883Z 2025-09-07T08:16:09.8059966Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8060042Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8060120Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8060188Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8060293Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8060489Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8060549Z return mod(*inputs) 2025-09-07T08:16:09.8060829Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8060898Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8061165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8061231Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8061454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8061536Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8061861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8061950Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8062211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8062346Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8062644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.8062771Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.8062775Z 2025-09-07T08:16:09.8062881Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8063075Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8063141Z return mod(*inputs) 2025-09-07T08:16:09.8063408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8063473Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8063741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8063841Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8064069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8064142Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8064494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8064580Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8064846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8064937Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8065229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.8065332Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.8065339Z 2025-09-07T08:16:09.8065433Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8065620Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8065677Z return mod(*inputs) 2025-09-07T08:16:09.8065943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8066007Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8066264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8066331Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8066558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8066631Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8066891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8066971Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8067231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.8067303Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.8067308Z 2025-09-07T08:16:09.8067403Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8067588Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8067688Z return mod(*inputs) 2025-09-07T08:16:09.8067951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8068013Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8068274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8068368Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8068583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8068650Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8068912Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8069025Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8069031Z 2025-09-07T08:16:09.8069121Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8069306Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8069361Z return mod(*inputs) 2025-09-07T08:16:09.8069657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8069724Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8069982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8070047Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8070306Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8070379Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8070638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8070742Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8070955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.8071021Z return self.act(input) 2025-09-07T08:16:09.8071026Z 2025-09-07T08:16:09.8071121Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8071309Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8071368Z return mod(*inputs) 2025-09-07T08:16:09.8071631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8071695Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8071953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8072016Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8072228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8072293Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8072553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.8072633Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.8072636Z 2025-09-07T08:16:09.8072724Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8072912Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8072968Z return mod(*inputs) 2025-09-07T08:16:09.8073233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8073302Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8073961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8074026Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8074248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8074354Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8074619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8074701Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8074968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.8075071Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.8075075Z 2025-09-07T08:16:09.8075168Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8075359Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8075415Z return mod(*inputs) 2025-09-07T08:16:09.8075681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8075780Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8076046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8076109Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8076323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8076434Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8076695Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8076784Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8077043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.8077207Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8077212Z 2025-09-07T08:16:09.8077305Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8077492Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8077550Z return mod(*inputs) 2025-09-07T08:16:09.8077813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8077881Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8078138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8078202Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8078415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8078481Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8078741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8078818Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8079081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.8079240Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8079244Z 2025-09-07T08:16:09.8079319Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8079390Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8079460Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8079570Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8079661Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8079849Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8079915Z return mod(*inputs) 2025-09-07T08:16:09.8080219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8080290Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8080549Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8080615Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8080837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8080906Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8081164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8081239Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8081498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8081620Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8081917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.8082046Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.8082049Z 2025-09-07T08:16:09.8082175Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8082363Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8082421Z return mod(*inputs) 2025-09-07T08:16:09.8082690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8082759Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8083017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8083084Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8083296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8083364Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8083621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8083698Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8083959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8084044Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8084339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.8084439Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.8084444Z 2025-09-07T08:16:09.8084535Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8084722Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8084777Z return mod(*inputs) 2025-09-07T08:16:09.8085038Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8085103Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8085361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8085459Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8085671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8085738Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8085998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8086111Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8086368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.8086438Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.8086442Z 2025-09-07T08:16:09.8086533Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8086714Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8086773Z return mod(*inputs) 2025-09-07T08:16:09.8087032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8087097Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8087396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8087463Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8087683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8087755Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8088055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:16:09.8088129Z hidden_states = residual + hidden_states 2025-09-07T08:16:09.8088132Z 2025-09-07T08:16:09.8088225Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8088417Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8088476Z return mod(*inputs) 2025-09-07T08:16:09.8088743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8088813Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8089075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8089145Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8089359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8089435Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8089693Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8089805Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8089808Z 2025-09-07T08:16:09.8089898Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8090082Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8090145Z return mod(*inputs) 2025-09-07T08:16:09.8090407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8090473Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8090733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8090799Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8091016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8091082Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8091389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8091492Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8091707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.8091802Z return self.act(input) 2025-09-07T08:16:09.8091805Z 2025-09-07T08:16:09.8091897Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8092086Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8092141Z return mod(*inputs) 2025-09-07T08:16:09.8092411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8092473Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8092731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8092795Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8093010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8093086Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8093378Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.8093452Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.8093459Z 2025-09-07T08:16:09.8093552Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8093786Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8093848Z return mod(*inputs) 2025-09-07T08:16:09.8094113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8094186Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8094446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8094508Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8094724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8094792Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8095053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8095133Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8095391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.8095494Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.8095500Z 2025-09-07T08:16:09.8095589Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8095780Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8095837Z return mod(*inputs) 2025-09-07T08:16:09.8096105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8096168Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8096428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8096496Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8096711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8096783Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8097082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8097162Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8097426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.8097657Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8097661Z 2025-09-07T08:16:09.8097758Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8097943Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8098005Z return mod(*inputs) 2025-09-07T08:16:09.8098268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8098330Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8098594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8098656Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8098869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8098938Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8099228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8099309Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8099569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.8099762Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8099766Z 2025-09-07T08:16:09.8099842Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8099920Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8099990Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8100059Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8100154Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8100340Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8100402Z return mod(*inputs) 2025-09-07T08:16:09.8100667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8100731Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8100993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8101054Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8101267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8101334Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8101590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8101667Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8101931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8102019Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8102313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.8102438Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.8102442Z 2025-09-07T08:16:09.8102532Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8102716Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8102810Z return mod(*inputs) 2025-09-07T08:16:09.8103251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8103318Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8103638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8103703Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8103921Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8103989Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8104253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8104332Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8104591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8104678Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8104968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.8105124Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.8105128Z 2025-09-07T08:16:09.8105218Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8105404Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8105459Z return mod(*inputs) 2025-09-07T08:16:09.8105776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8105845Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8106106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8106178Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8106388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8106458Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8106724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8106802Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8107067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.8107141Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.8107145Z 2025-09-07T08:16:09.8107238Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8107422Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8107480Z return mod(*inputs) 2025-09-07T08:16:09.8107744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8107811Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8108074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8108136Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8108347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8108420Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8108680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8108860Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8108863Z 2025-09-07T08:16:09.8108951Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8109138Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8109193Z return mod(*inputs) 2025-09-07T08:16:09.8109493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8109560Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8109816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8109881Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8110096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8110162Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8110425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8110529Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8110741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.8110844Z return self.act(input) 2025-09-07T08:16:09.8110847Z 2025-09-07T08:16:09.8110942Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8111127Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8111182Z return mod(*inputs) 2025-09-07T08:16:09.8111489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8111554Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8111816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8111882Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8112094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8112162Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8112422Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.8112496Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.8112499Z 2025-09-07T08:16:09.8112586Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8112770Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8112828Z return mod(*inputs) 2025-09-07T08:16:09.8113086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8113151Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8113414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8113481Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8113694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8113760Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8114021Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8114098Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8114361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.8114461Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.8114504Z 2025-09-07T08:16:09.8114594Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8114782Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8114837Z return mod(*inputs) 2025-09-07T08:16:09.8115103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8115199Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8115461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8115522Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8115735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8115806Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8116064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8116148Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8116405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.8116605Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8116609Z 2025-09-07T08:16:09.8116705Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8116888Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8116948Z return mod(*inputs) 2025-09-07T08:16:09.8117245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8117319Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8117579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8117645Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8117857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8117924Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8118189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8118266Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8118523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.8118681Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8118685Z 2025-09-07T08:16:09.8118757Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8118829Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8118896Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8118965Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8119057Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8119239Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8119300Z return mod(*inputs) 2025-09-07T08:16:09.8119558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8119622Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8119879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8119942Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8120157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8120264Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8120525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8120599Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8120858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8120989Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8121285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.8121416Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.8121421Z 2025-09-07T08:16:09.8121513Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8121706Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8121768Z return mod(*inputs) 2025-09-07T08:16:09.8122035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8122112Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8122406Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8122482Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8122698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8122772Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8123074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8123157Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8123426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8123511Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8123803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.8123911Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.8123915Z 2025-09-07T08:16:09.8124007Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8124196Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8124254Z return mod(*inputs) 2025-09-07T08:16:09.8124521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8124586Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8124846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8124915Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8125126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8125202Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8125463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8125541Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8125807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.8125882Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.8125885Z 2025-09-07T08:16:09.8125983Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8126207Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8126270Z return mod(*inputs) 2025-09-07T08:16:09.8126534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8126599Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8126902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8126966Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8127186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8127255Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8127516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:16:09.8127593Z hidden_states = residual + hidden_states 2025-09-07T08:16:09.8127598Z 2025-09-07T08:16:09.8127690Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8127881Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8127938Z return mod(*inputs) 2025-09-07T08:16:09.8128237Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8128304Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8128565Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8128634Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8128884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8128960Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8129221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8129331Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8129335Z 2025-09-07T08:16:09.8129434Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8129620Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8129684Z return mod(*inputs) 2025-09-07T08:16:09.8129948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8130013Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8130280Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8130345Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8130566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8130636Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8130901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8131004Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8131215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.8131286Z return self.act(input) 2025-09-07T08:16:09.8131290Z 2025-09-07T08:16:09.8131378Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8131570Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8131633Z return mod(*inputs) 2025-09-07T08:16:09.8131899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8132002Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8132263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8132335Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8132551Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8132657Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8132930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.8133008Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.8133012Z 2025-09-07T08:16:09.8133119Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8133311Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8133378Z return mod(*inputs) 2025-09-07T08:16:09.8133649Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8133717Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8133986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8134086Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8134314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8134386Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8134647Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8134771Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8135035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.8135149Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.8135153Z 2025-09-07T08:16:09.8135254Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8135456Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8135521Z return mod(*inputs) 2025-09-07T08:16:09.8135792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8135865Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8136126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8136198Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8136418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8136494Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8136760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8136843Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8137109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.8137269Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8137273Z 2025-09-07T08:16:09.8137376Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8137611Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8137674Z return mod(*inputs) 2025-09-07T08:16:09.8137946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8138047Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8138314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8138380Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8138600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8138725Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8138990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8139076Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8139338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.8139506Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8139512Z 2025-09-07T08:16:09.8139589Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8139659Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8139735Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8139803Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8139902Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8140152Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8140213Z return mod(*inputs) 2025-09-07T08:16:09.8140484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8140550Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8140855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8140922Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8141140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8141215Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8141475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8141563Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8141822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8141919Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8142219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.8142344Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.8142348Z 2025-09-07T08:16:09.8142448Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8142634Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8142699Z return mod(*inputs) 2025-09-07T08:16:09.8142962Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8143031Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8143298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8143361Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8143583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8143655Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8143922Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8144035Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8144295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8144390Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8144718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.8144826Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.8144830Z 2025-09-07T08:16:09.8144924Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8145113Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8145179Z return mod(*inputs) 2025-09-07T08:16:09.8145442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8145516Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8145777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8145849Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8146097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8146171Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8146437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8146517Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8146822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.8146898Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.8146903Z 2025-09-07T08:16:09.8146995Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8147187Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8147244Z return mod(*inputs) 2025-09-07T08:16:09.8147516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8147580Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8147843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8147906Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8148118Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8148187Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8148442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8148559Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8148563Z 2025-09-07T08:16:09.8148650Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8148831Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8148893Z return mod(*inputs) 2025-09-07T08:16:09.8149153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8149216Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8149478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8149540Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8149754Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8149861Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8150122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8150223Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8150476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.8150538Z return self.act(input) 2025-09-07T08:16:09.8150541Z 2025-09-07T08:16:09.8150630Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8150815Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8150871Z return mod(*inputs) 2025-09-07T08:16:09.8151135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8151198Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8151454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8151517Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8151728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8151829Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8152093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.8152167Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.8152171Z 2025-09-07T08:16:09.8152263Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8152486Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8152546Z return mod(*inputs) 2025-09-07T08:16:09.8152810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8152874Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8153130Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8153195Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8153407Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8153473Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8153732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8153812Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8154070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.8154175Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.8154179Z 2025-09-07T08:16:09.8154269Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8154454Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8154510Z return mod(*inputs) 2025-09-07T08:16:09.8154776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8154838Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8155094Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8155160Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8155371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8155475Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8155733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8155809Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8156071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.8156256Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8156260Z 2025-09-07T08:16:09.8156351Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8156533Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8156593Z return mod(*inputs) 2025-09-07T08:16:09.8156854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8156919Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8157177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8157237Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8157450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8157550Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8157808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8157890Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8158193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.8158352Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8158359Z 2025-09-07T08:16:09.8158431Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8158500Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8158567Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8158633Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8158723Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8158908Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8158968Z return mod(*inputs) 2025-09-07T08:16:09.8159229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8159290Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8159552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8159613Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8159827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8159893Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8160150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8160230Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8160486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8160573Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8160865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.8160989Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.8160992Z 2025-09-07T08:16:09.8161083Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8161320Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8161380Z return mod(*inputs) 2025-09-07T08:16:09.8161645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8161752Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8162012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8162075Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8162290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8162357Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8162617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8162695Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8162955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8163037Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8163356Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.8163460Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.8163464Z 2025-09-07T08:16:09.8163552Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8163739Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8163828Z return mod(*inputs) 2025-09-07T08:16:09.8164092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8164163Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8164421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8164488Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8164702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8164771Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8165027Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8165103Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8165363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.8165435Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.8165438Z 2025-09-07T08:16:09.8165531Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8165714Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8165770Z return mod(*inputs) 2025-09-07T08:16:09.8166035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8166101Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8166359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8166421Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8166633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8166703Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8166960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:16:09.8167067Z hidden_states = residual + hidden_states 2025-09-07T08:16:09.8167069Z 2025-09-07T08:16:09.8167158Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8167341Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8167430Z return mod(*inputs) 2025-09-07T08:16:09.8167693Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8167758Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8168025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8168092Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8168302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8168369Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8168633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8168738Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8168742Z 2025-09-07T08:16:09.8168833Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8169047Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8169107Z return mod(*inputs) 2025-09-07T08:16:09.8169369Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8169432Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8169722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8169785Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8170018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8170087Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8170364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8170477Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8170690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.8170755Z return self.act(input) 2025-09-07T08:16:09.8170758Z 2025-09-07T08:16:09.8170851Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8171048Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8171107Z return mod(*inputs) 2025-09-07T08:16:09.8171374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8171447Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8171707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8171778Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8171996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8172067Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8172336Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.8172413Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.8172417Z 2025-09-07T08:16:09.8172514Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8172703Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8172798Z return mod(*inputs) 2025-09-07T08:16:09.8173070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8173136Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8173443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8173509Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8173729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8173800Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8174059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8174149Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8174409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.8174522Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.8174526Z 2025-09-07T08:16:09.8174618Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8174840Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8174905Z return mod(*inputs) 2025-09-07T08:16:09.8175168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8175239Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8175530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8175604Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8175819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8175892Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8176163Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8176240Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8176509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.8176669Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8176672Z 2025-09-07T08:16:09.8176766Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8176959Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8177018Z return mod(*inputs) 2025-09-07T08:16:09.8177287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8177353Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8177681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8177749Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8177963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8178040Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8178297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8178381Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8178638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.8178833Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8178841Z 2025-09-07T08:16:09.8178917Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8178987Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8179061Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8179164Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8179254Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8179444Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8179500Z return mod(*inputs) 2025-09-07T08:16:09.8179771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8179838Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8180100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8180164Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8180376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8180446Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8180738Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8180821Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8181078Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8181167Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8181513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.8181638Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.8181643Z 2025-09-07T08:16:09.8181736Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8181920Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8181980Z return mod(*inputs) 2025-09-07T08:16:09.8182243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8182309Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8182572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8182637Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8182854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8182921Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8183182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8183260Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8183516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8183609Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8183897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.8183999Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.8184003Z 2025-09-07T08:16:09.8184090Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8184275Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8184334Z return mod(*inputs) 2025-09-07T08:16:09.8184635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8184703Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8184960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8185073Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8185289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8185356Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8185617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8185694Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8185959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.8186037Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.8186040Z 2025-09-07T08:16:09.8186129Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8186316Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8186373Z return mod(*inputs) 2025-09-07T08:16:09.8186672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8186736Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8186994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8187059Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8187305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8187377Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8187635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8187741Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8187748Z 2025-09-07T08:16:09.8187839Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8188025Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8188083Z return mod(*inputs) 2025-09-07T08:16:09.8188342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8188409Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8188668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8188729Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8188944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8189010Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8189272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8189376Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8189582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.8189652Z return self.act(input) 2025-09-07T08:16:09.8189655Z 2025-09-07T08:16:09.8189744Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8189934Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8189989Z return mod(*inputs) 2025-09-07T08:16:09.8190249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8190351Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8190611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8190682Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8190943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8191026Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8191293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.8191371Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.8191374Z 2025-09-07T08:16:09.8191479Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8191674Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8191742Z return mod(*inputs) 2025-09-07T08:16:09.8192007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8192074Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8192375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8192444Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8192670Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8192740Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8193039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8193124Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8193386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.8193495Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.8193498Z 2025-09-07T08:16:09.8193592Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8193784Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8193841Z return mod(*inputs) 2025-09-07T08:16:09.8194104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8194174Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8194435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8194506Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8194718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8194792Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8195052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8195132Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8195397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.8195554Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8195557Z 2025-09-07T08:16:09.8195655Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8195842Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8195900Z return mod(*inputs) 2025-09-07T08:16:09.8196161Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8196259Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8196526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8196590Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8196845Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8196914Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8197170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8197256Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8197516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.8197677Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8197682Z 2025-09-07T08:16:09.8197757Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8197826Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8197896Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8197966Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8198097Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8198285Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8198344Z return mod(*inputs) 2025-09-07T08:16:09.8198608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8198703Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8198974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8199045Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8199261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8199330Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8199589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8199674Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8199933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8200028Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8200324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.8200449Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.8200454Z 2025-09-07T08:16:09.8200552Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8200736Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8200794Z return mod(*inputs) 2025-09-07T08:16:09.8201057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8201131Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8201391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8201455Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8201672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8201738Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8202001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8202115Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8202375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8202466Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8202790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.8202902Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.8202906Z 2025-09-07T08:16:09.8202998Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8203340Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8203402Z return mod(*inputs) 2025-09-07T08:16:09.8203667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8203737Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8203999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8204070Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8204366Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8204438Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8204701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8204827Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8205093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.8205170Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.8205173Z 2025-09-07T08:16:09.8205267Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8205454Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8205511Z return mod(*inputs) 2025-09-07T08:16:09.8205790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8205862Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8206137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8206208Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8206433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8206513Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8206775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:16:09.8206852Z hidden_states = residual + hidden_states 2025-09-07T08:16:09.8206855Z 2025-09-07T08:16:09.8206956Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8207154Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8207222Z return mod(*inputs) 2025-09-07T08:16:09.8207489Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8207560Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8207822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8207890Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8208107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8208226Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8208488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8208603Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8208654Z 2025-09-07T08:16:09.8208758Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8208946Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8209005Z return mod(*inputs) 2025-09-07T08:16:09.8209273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8209336Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8209597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8209663Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8209877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8209949Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8210241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8210350Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8210562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.8210632Z return self.act(input) 2025-09-07T08:16:09.8210635Z 2025-09-07T08:16:09.8210758Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8210943Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8211007Z return mod(*inputs) 2025-09-07T08:16:09.8211270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8211336Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8211592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8211658Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8211879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8211947Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8212212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.8212284Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.8212287Z 2025-09-07T08:16:09.8212377Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8212562Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8212619Z return mod(*inputs) 2025-09-07T08:16:09.8212884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8212949Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8213211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8213273Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8213487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8213561Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8213821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8213945Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8214202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.8214311Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.8214346Z 2025-09-07T08:16:09.8214437Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8214621Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8214683Z return mod(*inputs) 2025-09-07T08:16:09.8214944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8215013Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8215271Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8215333Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8215548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8215615Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8215913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8215993Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8216252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.8216412Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8216416Z 2025-09-07T08:16:09.8216535Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8216722Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8216780Z return mod(*inputs) 2025-09-07T08:16:09.8217048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8217112Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8217374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8217449Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8217692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8217767Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8218025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8218103Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8218370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.8218533Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8218536Z 2025-09-07T08:16:09.8218612Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8218681Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8218757Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8218824Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8218913Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8219101Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8219162Z return mod(*inputs) 2025-09-07T08:16:09.8219429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8219494Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8219796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8219871Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8220087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8220197Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8220455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8220535Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8220796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8220886Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8221184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.8221310Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.8221313Z 2025-09-07T08:16:09.8221406Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8221592Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8221684Z return mod(*inputs) 2025-09-07T08:16:09.8221951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8222015Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8222275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8222374Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8222591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8222666Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8222926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8223009Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8223270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8223363Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8223654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.8223757Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.8223760Z 2025-09-07T08:16:09.8223859Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8224047Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8224107Z return mod(*inputs) 2025-09-07T08:16:09.8224374Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8224439Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8224704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8224774Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8224994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8225060Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8225327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8225407Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8225664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.8225773Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.8225776Z 2025-09-07T08:16:09.8225866Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8226055Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8226162Z return mod(*inputs) 2025-09-07T08:16:09.8226426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8226491Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8226752Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8226818Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8227033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8227102Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8227363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8227472Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8227478Z 2025-09-07T08:16:09.8227609Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8227796Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8227856Z return mod(*inputs) 2025-09-07T08:16:09.8228116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8228212Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8228479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8228545Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8228764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8228831Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8229091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8229196Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8229403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.8229473Z return self.act(input) 2025-09-07T08:16:09.8229476Z 2025-09-07T08:16:09.8229570Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8229758Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8229815Z return mod(*inputs) 2025-09-07T08:16:09.8230077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8230146Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8230404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8230477Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8230689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8230759Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8231018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.8231092Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.8231095Z 2025-09-07T08:16:09.8231190Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8231416Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8231474Z return mod(*inputs) 2025-09-07T08:16:09.8231735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8231880Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8232145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8232209Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8232428Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8232501Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8232760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8232846Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8233104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.8233210Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.8233214Z 2025-09-07T08:16:09.8233302Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8233521Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8233580Z return mod(*inputs) 2025-09-07T08:16:09.8233841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8233912Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8234205Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8234271Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8234486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8234551Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8234815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8234894Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8235156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.8235309Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8235313Z 2025-09-07T08:16:09.8235408Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8235591Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8235647Z return mod(*inputs) 2025-09-07T08:16:09.8235913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8235974Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8236235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8236298Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8236510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8236579Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8236842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8236924Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8237182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.8237373Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8237380Z 2025-09-07T08:16:09.8237451Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8237518Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8237623Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8237692Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8237784Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8237987Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8238045Z return mod(*inputs) 2025-09-07T08:16:09.8238314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8238376Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8238639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8238704Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8238919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8238993Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8239287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8239371Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8239630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8239749Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8240050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.8240175Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.8240178Z 2025-09-07T08:16:09.8240276Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8240461Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8240525Z return mod(*inputs) 2025-09-07T08:16:09.8240789Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8240856Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8241122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8241189Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8241408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8241476Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8241735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8241819Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8242084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8242176Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8242463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.8242566Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.8242568Z 2025-09-07T08:16:09.8242661Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8242853Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8242955Z return mod(*inputs) 2025-09-07T08:16:09.8243218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8243286Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8243546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8243645Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8243863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8243929Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8244197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8244274Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8244541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.8244615Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.8244618Z 2025-09-07T08:16:09.8244706Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8244897Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8244988Z return mod(*inputs) 2025-09-07T08:16:09.8245262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8245327Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8245584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8245681Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8245895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8245965Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8246223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:16:09.8246294Z hidden_states = residual + hidden_states 2025-09-07T08:16:09.8246299Z 2025-09-07T08:16:09.8246392Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8246575Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8246637Z return mod(*inputs) 2025-09-07T08:16:09.8246897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8246967Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8247317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8247415Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8247676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8253879Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8254186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8254320Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8254325Z 2025-09-07T08:16:09.8254427Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8254644Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8254706Z return mod(*inputs) 2025-09-07T08:16:09.8255001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8255078Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8255424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8255496Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8255721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8255840Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8256113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8256231Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8256450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.8256518Z return self.act(input) 2025-09-07T08:16:09.8256522Z 2025-09-07T08:16:09.8256629Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8256833Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8256892Z return mod(*inputs) 2025-09-07T08:16:09.8257165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8257234Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8257620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8257688Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8257909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8257988Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8258290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.8258371Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.8258378Z 2025-09-07T08:16:09.8258474Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8258670Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8258730Z return mod(*inputs) 2025-09-07T08:16:09.8258994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8259064Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8259324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8259398Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8259625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8259702Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8259976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8260065Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8260329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.8260440Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.8260444Z 2025-09-07T08:16:09.8260549Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8260746Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8260807Z return mod(*inputs) 2025-09-07T08:16:09.8261081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8261151Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8261419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8261523Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8261743Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8261819Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8262118Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8262206Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8262475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.8262649Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8262655Z 2025-09-07T08:16:09.8262753Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8262945Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8263010Z return mod(*inputs) 2025-09-07T08:16:09.8263275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8263345Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8263639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8263708Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8263927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8263997Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8264292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8264372Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8264632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.8264792Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8264796Z 2025-09-07T08:16:09.8264871Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8264944Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8265012Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8265082Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8265174Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8265362Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8265423Z return mod(*inputs) 2025-09-07T08:16:09.8265698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8265769Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8266042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8266109Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8266337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8266410Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8266676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8266755Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8267014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8267108Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8267442Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.8267572Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.8267576Z 2025-09-07T08:16:09.8267674Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8267926Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8267985Z return mod(*inputs) 2025-09-07T08:16:09.8268251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8268321Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8268582Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8268648Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8268864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8268936Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8269194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8269275Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8269567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8269654Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8269945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.8270077Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.8270081Z 2025-09-07T08:16:09.8270173Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8270370Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8270429Z return mod(*inputs) 2025-09-07T08:16:09.8270694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8270758Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8271018Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8271085Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8271298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8271374Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8271632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8271720Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8271981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.8272053Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.8272056Z 2025-09-07T08:16:09.8272152Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8272341Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8272402Z return mod(*inputs) 2025-09-07T08:16:09.8272664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8272731Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8272992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8273057Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8273313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8273384Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8273641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8273796Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8273799Z 2025-09-07T08:16:09.8273891Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8274082Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8274141Z return mod(*inputs) 2025-09-07T08:16:09.8274423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8274491Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8274765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8274833Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8275050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8275122Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8275414Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8275524Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8275737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.8275833Z return self.act(input) 2025-09-07T08:16:09.8275837Z 2025-09-07T08:16:09.8275935Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8276130Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8276192Z return mod(*inputs) 2025-09-07T08:16:09.8276455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8276519Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8276781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8276847Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8277061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8277127Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8277384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.8277460Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.8277465Z 2025-09-07T08:16:09.8277557Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8277744Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8277805Z return mod(*inputs) 2025-09-07T08:16:09.8278065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8278136Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8278397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8278467Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8278682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8278754Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8279015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8279133Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8279398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.8279507Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.8279545Z 2025-09-07T08:16:09.8279645Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8279832Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8279892Z return mod(*inputs) 2025-09-07T08:16:09.8280162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8280230Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8280502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8280570Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8280790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8280860Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8281156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8281243Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8281504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.8281698Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8281703Z 2025-09-07T08:16:09.8281796Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8281983Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8282050Z return mod(*inputs) 2025-09-07T08:16:09.8282314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8282379Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8282639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8282704Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8282916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8282982Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8283242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8283318Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8283578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.8283731Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8283735Z 2025-09-07T08:16:09.8283807Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8283881Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8283950Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8284020Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8284108Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8284289Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8284349Z return mod(*inputs) 2025-09-07T08:16:09.8284615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8284719Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8284977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8285046Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8285261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8285364Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8285625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8285704Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8285972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8286063Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8286359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.8286490Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.8286493Z 2025-09-07T08:16:09.8286586Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8286807Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8286866Z return mod(*inputs) 2025-09-07T08:16:09.8287132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8287198Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8287488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8287559Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8287773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8287849Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8288109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8288186Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8288451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8288536Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8288830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.8288930Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.8288934Z 2025-09-07T08:16:09.8289029Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8289213Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8289273Z return mod(*inputs) 2025-09-07T08:16:09.8289536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8289600Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8289867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8289931Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8290143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8290214Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8290476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8290560Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8290870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.8290942Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.8290949Z 2025-09-07T08:16:09.8291038Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8291258Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8291324Z return mod(*inputs) 2025-09-07T08:16:09.8291585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8291654Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8291915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8291981Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8292199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8292267Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8292533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:16:09.8292607Z hidden_states = residual + hidden_states 2025-09-07T08:16:09.8292643Z 2025-09-07T08:16:09.8292735Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8292921Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8292980Z return mod(*inputs) 2025-09-07T08:16:09.8293275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8293342Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8293610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8293679Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8293891Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8293964Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8294226Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8294337Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8294341Z 2025-09-07T08:16:09.8294430Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8294612Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8294679Z return mod(*inputs) 2025-09-07T08:16:09.8294943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8295015Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8295273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8295341Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8295557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8295624Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8295885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8295988Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8296203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.8296267Z return self.act(input) 2025-09-07T08:16:09.8296270Z 2025-09-07T08:16:09.8296396Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8296586Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8296643Z return mod(*inputs) 2025-09-07T08:16:09.8296911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8297010Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8297268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8297330Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8297584Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8297655Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8297919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.8297997Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.8298001Z 2025-09-07T08:16:09.8298094Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8298279Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8298339Z return mod(*inputs) 2025-09-07T08:16:09.8298642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8298709Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8298976Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8299073Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8299290Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8299366Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8299624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8299707Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8299968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.8300071Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.8300080Z 2025-09-07T08:16:09.8300173Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8300357Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8300420Z return mod(*inputs) 2025-09-07T08:16:09.8300682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8300756Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8301015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8301081Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8301298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8301368Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8301634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8301712Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8301971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.8302132Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8302135Z 2025-09-07T08:16:09.8302260Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8302450Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8302507Z return mod(*inputs) 2025-09-07T08:16:09.8302775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8303254Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8303516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8303583Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8303799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8303872Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8304133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8304214Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8304478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.8304635Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8304704Z 2025-09-07T08:16:09.8304785Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8304853Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8304928Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8305002Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8305096Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8305334Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8305396Z return mod(*inputs) 2025-09-07T08:16:09.8305663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8305735Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8305998Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8306071Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8306289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8306364Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8306623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8306709Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8306975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8307066Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8307363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.8307488Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.8307493Z 2025-09-07T08:16:09.8307596Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8307780Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8307840Z return mod(*inputs) 2025-09-07T08:16:09.8308110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8308177Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8308439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8308556Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8308772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8308851Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8309114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8309239Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8309502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8309585Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8309883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.8309986Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.8309990Z 2025-09-07T08:16:09.8310087Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8310271Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8310332Z return mod(*inputs) 2025-09-07T08:16:09.8310633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8310701Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8310965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8311030Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8311291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8311362Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8311622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8311708Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8311964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.8312040Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.8312045Z 2025-09-07T08:16:09.8312137Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8312323Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8312379Z return mod(*inputs) 2025-09-07T08:16:09.8312640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8312706Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8312962Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8313034Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8313247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8313315Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8313581Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8313692Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8313695Z 2025-09-07T08:16:09.8313797Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8313981Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8314044Z return mod(*inputs) 2025-09-07T08:16:09.8314307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8314405Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8314671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8314738Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8314957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8315054Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8315315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8315423Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8315633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.8315699Z return self.act(input) 2025-09-07T08:16:09.8315703Z 2025-09-07T08:16:09.8315793Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8315985Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8316043Z return mod(*inputs) 2025-09-07T08:16:09.8316304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8316409Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8316673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8316742Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8316957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8317055Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8317319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.8317397Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.8317401Z 2025-09-07T08:16:09.8317494Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8317679Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8317737Z return mod(*inputs) 2025-09-07T08:16:09.8318007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8318071Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8318332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8318394Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8318614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8318683Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8318946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8319028Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8319287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 315, in forward 2025-09-07T08:16:09.8319396Z query_states = self.q_proj(hidden_states) * self.scaling 2025-09-07T08:16:09.8319400Z 2025-09-07T08:16:09.8319491Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8319674Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8319735Z return mod(*inputs) 2025-09-07T08:16:09.8319996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8320065Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8320358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8320427Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8320639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8320745Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8321010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8321090Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8321352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 336, in forward 2025-09-07T08:16:09.8321506Z key_states = self.k_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8321511Z 2025-09-07T08:16:09.8321601Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8321797Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8321859Z return mod(*inputs) 2025-09-07T08:16:09.8322136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8322268Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8322538Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8322607Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8322822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8322930Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8323193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8323283Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8323544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 337, in forward 2025-09-07T08:16:09.8323705Z value_states = self.v_proj(current_states).view(bsz, -1, self.num_heads, self.head_dim) 2025-09-07T08:16:09.8323710Z 2025-09-07T08:16:09.8323793Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8323862Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8323937Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8324006Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8324100Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8324290Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8324354Z return mod(*inputs) 2025-09-07T08:16:09.8324620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8324685Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8324946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8325013Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8325229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8325309Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8325568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8325652Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8325909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8325997Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8326334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 81, in sdpa_attention_forward 2025-09-07T08:16:09.8326461Z attn_output = torch.nn.functional.scaled_dot_product_attention( 2025-09-07T08:16:09.8326465Z 2025-09-07T08:16:09.8326564Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8326788Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8326854Z return mod(*inputs) 2025-09-07T08:16:09.8327115Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8327180Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8327447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8327515Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8327735Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8327804Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8328062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8328182Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8328440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 351, in forward 2025-09-07T08:16:09.8328530Z attn_output, attn_weights = attention_interface( 2025-09-07T08:16:09.8328852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/integrations/sdpa_attention.py", line 91, in sdpa_attention_forward 2025-09-07T08:16:09.8328955Z attn_output = attn_output.transpose(1, 2).contiguous() 2025-09-07T08:16:09.8328964Z 2025-09-07T08:16:09.8329059Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8329246Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8329311Z return mod(*inputs) 2025-09-07T08:16:09.8329572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8329643Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8329902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8329968Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8330190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8330260Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8330521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 410, in forward 2025-09-07T08:16:09.8330604Z hidden_states, attn_weights = self.self_attn( 2025-09-07T08:16:09.8330863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 365, in forward 2025-09-07T08:16:09.8330941Z attn_output = self.out_proj(attn_output) 2025-09-07T08:16:09.8330946Z 2025-09-07T08:16:09.8331037Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8331226Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8331283Z return mod(*inputs) 2025-09-07T08:16:09.8331549Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8331613Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8331872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8331977Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8332191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8332266Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8332527Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 417, in forward 2025-09-07T08:16:09.8332633Z hidden_states = residual + hidden_states 2025-09-07T08:16:09.8332636Z 2025-09-07T08:16:09.8332732Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8332915Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8332977Z return mod(*inputs) 2025-09-07T08:16:09.8333240Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8333308Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8333567Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8333631Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8333855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8333973Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8334241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8334354Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8334357Z 2025-09-07T08:16:09.8334447Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8334665Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8334723Z return mod(*inputs) 2025-09-07T08:16:09.8334993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8335059Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8335319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8335392Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8335607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8335678Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8335938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 421, in forward 2025-09-07T08:16:09.8336046Z hidden_states = self.activation_fn(self.fc1(hidden_states)) 2025-09-07T08:16:09.8336253Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T08:16:09.8336318Z return self.act(input) 2025-09-07T08:16:09.8336321Z 2025-09-07T08:16:09.8336421Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8336604Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8336667Z return mod(*inputs) 2025-09-07T08:16:09.8336931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1596, in forward 2025-09-07T08:16:09.8336996Z encoder_outputs = self.encoder( 2025-09-07T08:16:09.8337263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 708, in forward 2025-09-07T08:16:09.8337326Z layer_outputs = encoder_layer( 2025-09-07T08:16:09.8337613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T08:16:09.8337683Z return super().__call__(*args, **kwargs) 2025-09-07T08:16:09.8337983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 423, in forward 2025-09-07T08:16:09.8338059Z hidden_states = self.fc2(hidden_states) 2025-09-07T08:16:09.8338063Z 2025-09-07T08:16:09.8338154Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8338348Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8338439Z return mod(*inputs) 2025-09-07T08:16:09.8338705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1612, in forward 2025-09-07T08:16:09.8338785Z hidden_states = self.projector(hidden_states) 2025-09-07T08:16:09.8338788Z 2025-09-07T08:16:09.8338863Z cudagraph partition due to non gpu ops 2025-09-07T08:16:09.8338955Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8339138Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8339201Z return mod(*inputs) 2025-09-07T08:16:09.8339459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1613, in forward 2025-09-07T08:16:09.8339539Z pooled_output = hidden_states.mean(dim=1) 2025-09-07T08:16:09.8339549Z 2025-09-07T08:16:09.8339675Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:09.8339860Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:09.8339922Z return mod(*inputs) 2025-09-07T08:16:09.8340181Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/whisper/modeling_whisper.py", line 1615, in forward 2025-09-07T08:16:09.8340263Z logits = self.classifier(pooled_output) 2025-09-07T08:16:09.8340297Z 2025-09-07T08:16:20.4040470Z pass 2025-09-07T08:16:20.4040879Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:16:23.2809711Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:16:23.2810613Z import pynvml # type: ignore[import] 2025-09-07T08:16:25.2933515Z 2025-09-07T08:16:25.4455816Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:16:25.4456083Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:16:25.4458158Z cpu eval lennard_jones 2025-09-07T08:16:25.4471091Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:16:25.4494203Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:16:25.4509529Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:16:27.3264490Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:27.3264939Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:27.3265323Z return mod(*inputs) 2025-09-07T08:16:27.3265435Z 2025-09-07T08:16:27.3265558Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:27.3265962Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:27.3266292Z return mod(*inputs) 2025-09-07T08:16:27.3266403Z 2025-09-07T08:16:27.3266505Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:27.3266885Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:27.3267231Z return mod(*inputs) 2025-09-07T08:16:27.3267361Z 2025-09-07T08:16:27.3267466Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:27.3267830Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:27.3268501Z return mod(*inputs) 2025-09-07T08:16:27.3268598Z 2025-09-07T08:16:27.3268704Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:27.3269049Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:27.3269471Z return mod(*inputs) 2025-09-07T08:16:27.3269571Z 2025-09-07T08:16:27.3269668Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:27.3270013Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:27.3270328Z return mod(*inputs) 2025-09-07T08:16:27.3270431Z 2025-09-07T08:16:27.3270531Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:27.3270891Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:27.3271217Z return mod(*inputs) 2025-09-07T08:16:27.3271311Z 2025-09-07T08:16:27.3271414Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:27.3271750Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:27.3272071Z return mod(*inputs) 2025-09-07T08:16:27.3272165Z 2025-09-07T08:16:27.3272261Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T08:16:27.3272709Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:16:27.3273040Z return mod(*inputs) 2025-09-07T08:16:27.3273139Z 2025-09-07T08:16:31.5218480Z pass 2025-09-07T08:16:31.5220260Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:16:33.0244492Z accuracy pass_rate=91.30% 2025-09-07T08:16:33.0248602Z calls_captured gmean=0.00x mean=416.435x 2025-09-07T08:16:33.0251255Z unique_graphs gmean=0.00x mean=3.000x 2025-09-07T08:16:33.0253533Z graph_breaks gmean=0.00x mean=2.174x 2025-09-07T08:16:33.0255772Z unique_graph_breaks gmean=0.00x mean=0.565x 2025-09-07T08:16:33.0257977Z autograd_captures gmean=0.00x mean=0.000x 2025-09-07T08:16:33.0260101Z autograd_compiles gmean=0.00x mean=0.000x 2025-09-07T08:16:33.0262173Z cudagraph_skips gmean=0.00x mean=2.435x 2025-09-07T08:16:33.0263033Z compilation_latency mean=23.057 seconds 2025-09-07T08:16:33.5462490Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *cppwrapper-true* ]] 2025-09-07T08:16:33.5462991Z + TORCHINDUCTOR_CPP_WRAPPER=1 2025-09-07T08:16:33.5463953Z + taskset -c 0-94 python benchmarks/dynamo/torchbench.py --accuracy --no-translation-validation --inference --bfloat16 --backend inductor --disable-cudagraphs --device cpu --total-partitions 4 --partition-id 1 --output /var/lib/jenkins/workspace/test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.csv 2025-09-07T08:16:33.9420744Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:16:33.9421596Z import pynvml # type: ignore[import] 2025-09-07T08:16:36.3836622Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:16:36.3837521Z import pynvml # type: ignore[import] 2025-09-07T08:16:38.4053482Z 2025-09-07T08:16:43.6439312Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:16:43.6439649Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:16:43.6720873Z cpu eval dlrm 2025-09-07T08:16:43.7998485Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:16:43.8246854Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:16:43.8503872Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:16:55.0486633Z pass 2025-09-07T08:16:55.0488622Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:16:57.2445476Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:16:57.2446604Z import pynvml # type: ignore[import] 2025-09-07T08:16:59.2551329Z 2025-09-07T08:17:01.9431538Z loading model: 0it [00:00, ?it/s]WARNING:common:Model doctr_det_predictor does not support bfloat16, running with amp instead 2025-09-07T08:17:02.5709082Z 2025-09-07T08:17:02.5709607Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:17:02.5710011Z WARNING:common:Model doctr_det_predictor does not support bfloat16, running with amp instead 2025-09-07T08:17:02.5710414Z cpu eval doctr_det_predictor 2025-09-07T08:17:17.1846392Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:17:17.1847326Z WARNING:common:Model doctr_det_predictor does not support bfloat16, running with amp instead 2025-09-07T08:17:18.0914109Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:17:18.9938880Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:17:50.1677520Z pass 2025-09-07T08:17:50.1682250Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:17:53.4743687Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:17:53.4744701Z import pynvml # type: ignore[import] 2025-09-07T08:17:55.4822340Z 2025-09-07T08:17:58.1519783Z loading model: 0it [00:00, ?it/s]WARNING:common:Model doctr_reco_predictor does not support bfloat16, running with amp instead 2025-09-07T08:17:58.1956444Z 2025-09-07T08:17:58.1956785Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:17:58.1957216Z WARNING:common:Model doctr_reco_predictor does not support bfloat16, running with amp instead 2025-09-07T08:17:58.1958086Z cpu eval doctr_reco_predictor 2025-09-07T08:17:58.6806269Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:17:58.6806841Z WARNING:common:Model doctr_reco_predictor does not support bfloat16, running with amp instead 2025-09-07T08:17:58.7417769Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:17:58.7898380Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:11.7831695Z pass 2025-09-07T08:18:11.7836176Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:14.0377732Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:18:14.0378618Z import pynvml # type: ignore[import] 2025-09-07T08:18:16.0371641Z 2025-09-07T08:18:16.0898904Z loading model: 0it [00:00, ?it/s]Gym has been unmaintained since 2022 and does not support NumPy 2.0 amongst other critical functionality. 2025-09-07T08:18:16.0899973Z Please upgrade to Gymnasium, the maintained drop-in replacement of Gym, or contact the authors of your software and request that they upgrade. 2025-09-07T08:18:16.0900582Z Users of this version of Gym should be able to simply replace 'import gym' with 'import gymnasium as gym' in the vast majority of cases. 2025-09-07T08:18:16.0901175Z See the migration guide at https://gymnasium.farama.org/introduction/migration_guide/ for additional information. 2025-09-07T08:18:16.9767255Z 2025-09-07T08:18:16.9767691Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:18:16.9782752Z cpu eval drq 2025-09-07T08:18:16.9958442Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:17.0061529Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:17.0135689Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:26.0446735Z pass 2025-09-07T08:18:26.0447131Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:28.0956929Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:18:28.0957841Z import pynvml # type: ignore[import] 2025-09-07T08:18:30.1174330Z 2025-09-07T08:18:31.7243720Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:18:31.7244039Z loading model: 0it [00:01, ?it/s] 2025-09-07T08:18:31.7427205Z cpu eval fastNLP_Bert 2025-09-07T08:18:32.7436864Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:33.1387317Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:33.5262712Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:33.6140962Z W0907 08:18:33.613451 68348 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] Graph break from `Tensor.item()`, consider setting: 2025-09-07T08:18:33.6141679Z W0907 08:18:33.613451 68348 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] torch._dynamo.config.capture_scalar_outputs = True 2025-09-07T08:18:33.6142281Z W0907 08:18:33.613451 68348 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] or: 2025-09-07T08:18:33.6142764Z W0907 08:18:33.613451 68348 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1 2025-09-07T08:18:33.6143351Z W0907 08:18:33.613451 68348 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] to include these operations in the captured graph. 2025-09-07T08:18:33.6143833Z W0907 08:18:33.613451 68348 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:18:33.6144295Z W0907 08:18:33.613451 68348 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] Graph break: from user code at: 2025-09-07T08:18:33.6144970Z W0907 08:18:33.613451 68348 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 482, in forward_pass 2025-09-07T08:18:33.6145624Z W0907 08:18:33.613451 68348 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] return mod(*inputs) 2025-09-07T08:18:33.6146273Z W0907 08:18:33.613451 68348 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/models/bert.py", line 265, in forward 2025-09-07T08:18:33.6146946Z W0907 08:18:33.613451 68348 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] sequence_output = self.bert(words) 2025-09-07T08:18:33.6147657Z W0907 08:18:33.613451 68348 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 137, in forward 2025-09-07T08:18:33.6148702Z W0907 08:18:33.613451 68348 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] outputs = self.model(words) 2025-09-07T08:18:33.6149408Z W0907 08:18:33.613451 68348 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 445, in forward 2025-09-07T08:18:33.6150770Z W0907 08:18:33.613451 68348 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] max_word_piece_length = batch_word_pieces_length.sum(dim=-1).max().item() # 表示word piece的长度(包括padding) 2025-09-07T08:18:33.6151367Z W0907 08:18:33.613451 68348 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:18:33.6151750Z W0907 08:18:33.613451 68348 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:18:53.5345870Z pass 2025-09-07T08:18:53.5349208Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:56.0726570Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:18:56.0727697Z import pynvml # type: ignore[import] 2025-09-07T08:18:58.0833490Z 2025-09-07T08:18:58.5146635Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:18:58.5146935Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:18:58.5180141Z cpu eval functorch_dp_cifar10 2025-09-07T08:18:58.5854916Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:58.6124178Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:18:58.6361458Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:19:11.7422220Z pass 2025-09-07T08:19:11.7424584Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:19:13.9312655Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:19:13.9313827Z import pynvml # type: ignore[import] 2025-09-07T08:19:15.9435138Z 2025-09-07T08:19:16.1359876Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:19:16.1360169Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:19:16.1360438Z cpu eval functorch_maml_omniglot 2025-09-07T08:19:16.1464147Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:19:16.1541375Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:19:16.1588731Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:19:24.6099789Z pass 2025-09-07T08:19:24.6102075Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:19:26.5856713Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:19:26.5857631Z import pynvml # type: ignore[import] 2025-09-07T08:19:28.6000579Z 2025-09-07T08:19:30.2791902Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:19:30.2792377Z loading model: 0it [00:01, ?it/s] 2025-09-07T08:19:30.2842126Z cpu eval hf_Albert 2025-09-07T08:19:31.2003558Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:19:31.4338149Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:19:31.7299460Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:19:46.7909985Z pass 2025-09-07T08:19:46.7910679Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:19:49.3119793Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:19:49.3120653Z import pynvml # type: ignore[import] 2025-09-07T08:19:51.3269898Z 2025-09-07T08:19:55.4931451Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:19:55.4932707Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:19:55.5159829Z cpu eval hf_Bart 2025-09-07T08:19:56.9244223Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:19:57.2758791Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:19:57.6244860Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:20:19.5499114Z pass 2025-09-07T08:20:19.5499514Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:20:22.4469439Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:20:22.4470326Z import pynvml # type: ignore[import] 2025-09-07T08:20:24.4715413Z 2025-09-07T08:20:27.5908612Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:20:27.5908928Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:20:27.6145578Z cpu eval hf_Bert 2025-09-07T08:20:28.7199638Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:20:29.0044966Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:20:29.2825978Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:20:46.3649530Z pass 2025-09-07T08:20:46.3649936Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:20:49.1225631Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:20:49.1226499Z import pynvml # type: ignore[import] 2025-09-07T08:20:51.1318573Z 2025-09-07T08:20:56.4991772Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:20:56.4992094Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:20:56.5398449Z cpu eval hf_Bert_large 2025-09-07T08:20:58.8700592Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:20:59.5192138Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:21:00.1590353Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:21:28.6826724Z pass 2025-09-07T08:21:28.6827123Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:21:31.7273732Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:21:31.7275073Z import pynvml # type: ignore[import] 2025-09-07T08:21:33.7340128Z 2025-09-07T08:21:35.1280850Z loading model: 0it [00:00, ?it/s]BigBirdForMaskedLM has generative capabilities, as `prepare_inputs_for_generation` is explicitly defined. However, it doesn't directly inherit from `GenerationMixin`. From 👉v4.50👈 onwards, `PreTrainedModel` will NOT inherit from `GenerationMixin`, and this model will lose the ability to call `generate` and other related functions. 2025-09-07T08:21:35.1282185Z - If you're using `trust_remote_code=True`, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes 2025-09-07T08:21:35.1283004Z - If you are the owner of the model architecture code, please modify your model class such that it inherits from `GenerationMixin` (after `PreTrainedModel`, otherwise you'll get an exception). 2025-09-07T08:21:35.1283659Z - If you are not the owner of the model architecture class, please contact the model code owner to update it. 2025-09-07T08:21:35.1285309Z WARNING:transformers.modeling_utils:BigBirdForMaskedLM has generative capabilities, as `prepare_inputs_for_generation` is explicitly defined. However, it doesn't directly inherit from `GenerationMixin`. From 👉v4.50👈 onwards, `PreTrainedModel` will NOT inherit from `GenerationMixin`, and this model will lose the ability to call `generate` and other related functions. 2025-09-07T08:21:35.1286687Z - If you're using `trust_remote_code=True`, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes 2025-09-07T08:21:35.1287512Z - If you are the owner of the model architecture code, please modify your model class such that it inherits from `GenerationMixin` (after `PreTrainedModel`, otherwise you'll get an exception). 2025-09-07T08:21:35.1288155Z - If you are not the owner of the model architecture class, please contact the model code owner to update it. 2025-09-07T08:21:36.9583114Z Input ids are automatically padded from 819 to 832 to be a multiple of `config.block_size`: 64 2025-09-07T08:21:36.9583922Z WARNING:transformers.models.big_bird.modeling_big_bird:Input ids are automatically padded from 819 to 832 to be a multiple of `config.block_size`: 64 2025-09-07T08:21:38.8402477Z 2025-09-07T08:21:38.8403809Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:21:38.8632969Z cpu eval hf_BigBird 2025-09-07T08:21:41.3858395Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:21:44.4333237Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:21:47.4215069Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:22:18.3015505Z W0907 08:22:18.300029 76432 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] torch._dynamo hit config.recompile_limit (8) 2025-09-07T08:22:18.3016461Z W0907 08:22:18.300029 76432 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] function: 'forward' (/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py:417) 2025-09-07T08:22:18.3018379Z W0907 08:22:18.300029 76432 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] last reason: 10/7: self.seed == 7 # context_layer, attention_probs = self.bigbird_block_sparse_attention( # transformers/models/big_bird/modeling_big_bird.py:455 in forward (HINT: torch.compile considers integer attributes of the nn.Module to be static. If you are observing recompilation, you might want to make this integer dynamic using torch._dynamo.config.allow_unspec_int_on_nn_module = True, or convert this integer into a tensor.) 2025-09-07T08:22:18.3020489Z W0907 08:22:18.300029 76432 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] To log all recompilation reasons, use TORCH_LOGS="recompiles". 2025-09-07T08:22:18.3021282Z W0907 08:22:18.300029 76432 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] To diagnose recompilation issues, see https://pytorch.org/docs/main/torch.compiler_troubleshooting.html 2025-09-07T08:22:23.7943515Z pass 2025-09-07T08:22:23.7943902Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:22:26.9062239Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:22:26.9063148Z import pynvml # type: ignore[import] 2025-09-07T08:22:28.9265988Z 2025-09-07T08:22:31.5187188Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:22:31.5187522Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:22:31.5318948Z cpu eval hf_DistilBert 2025-09-07T08:22:32.1196178Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:22:32.2821009Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:22:32.4345785Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:22:44.8162669Z pass 2025-09-07T08:22:44.8163072Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:22:47.3707842Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:22:47.3708746Z import pynvml # type: ignore[import] 2025-09-07T08:22:49.3874184Z 2025-09-07T08:22:54.0575934Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:22:54.0576334Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:22:54.0752338Z cpu eval hf_GPT2 2025-09-07T08:22:54.6321179Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:22:54.7957659Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:22:54.9599357Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:23:12.1194164Z pass 2025-09-07T08:23:12.1194588Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:23:14.8965698Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:23:14.8966726Z import pynvml # type: ignore[import] 2025-09-07T08:23:16.9138560Z 2025-09-07T08:23:31.1541606Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:23:31.1541942Z loading model: 0it [00:14, ?it/s] 2025-09-07T08:23:31.2149438Z cpu eval hf_GPT2_large 2025-09-07T08:23:31.6070923Z pass_due_to_skip 2025-09-07T08:23:31.6075261Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:23:34.0089216Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:23:34.0090412Z import pynvml # type: ignore[import] 2025-09-07T08:23:36.0233091Z 2025-09-07T08:23:40.0757813Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:23:40.0758107Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:23:40.1051988Z cpu eval hf_Longformer 2025-09-07T08:23:41.8437702Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:23:42.6335270Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:23:43.3739727Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:23:43.5444359Z W0907 08:23:43.543764 81358 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] Graph break from `Tensor.item()`, consider setting: 2025-09-07T08:23:43.5445021Z W0907 08:23:43.543764 81358 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] torch._dynamo.config.capture_scalar_outputs = True 2025-09-07T08:23:43.5445548Z W0907 08:23:43.543764 81358 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] or: 2025-09-07T08:23:43.5446054Z W0907 08:23:43.543764 81358 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1 2025-09-07T08:23:43.5447018Z W0907 08:23:43.543764 81358 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] to include these operations in the captured graph. 2025-09-07T08:23:43.5447513Z W0907 08:23:43.543764 81358 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:23:43.5447964Z W0907 08:23:43.543764 81358 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] Graph break: from user code at: 2025-09-07T08:23:43.5448728Z W0907 08:23:43.543764 81358 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/var/lib/jenkins/workspace/benchmarks/dynamo/torchbench.py", line 480, in forward_pass 2025-09-07T08:23:43.5449401Z W0907 08:23:43.543764 81358 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] return mod(**inputs) 2025-09-07T08:23:43.5450174Z W0907 08:23:43.543764 81358 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1703, in forward 2025-09-07T08:23:43.5450938Z W0907 08:23:43.543764 81358 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] outputs = self.longformer( 2025-09-07T08:23:43.5451726Z W0907 08:23:43.543764 81358 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1600, in forward 2025-09-07T08:23:43.5452484Z W0907 08:23:43.543764 81358 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] encoder_outputs = self.encoder( 2025-09-07T08:23:43.5453223Z W0907 08:23:43.543764 81358 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1244, in forward 2025-09-07T08:23:43.5454018Z W0907 08:23:43.543764 81358 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] is_global_attn = is_index_global_attn.flatten().any().item() 2025-09-07T08:23:43.5454512Z W0907 08:23:43.543764 81358 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:23:43.5454881Z W0907 08:23:43.543764 81358 site-packages/torch/_dynamo/variables/tensor.py:1048] [0/0] 2025-09-07T08:25:23.7218915Z pass 2025-09-07T08:25:23.7219309Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:27.7222502Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:25:27.7223729Z import pynvml # type: ignore[import] 2025-09-07T08:25:29.7373845Z 2025-09-07T08:25:31.1147093Z loading model: 0it [00:00, ?it/s]Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T08:25:31.1149437Z WARNING:transformers.models.reformer.modeling_reformer:Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T08:25:31.8694205Z 2025-09-07T08:25:31.8694624Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:25:31.8744925Z cpu eval hf_Reformer 2025-09-07T08:25:32.4035723Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:33.8553986Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:35.3094055Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:25:42.5542031Z W0907 08:25:42.553452 82806 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] Graph break from `Tensor.item()`, consider setting: 2025-09-07T08:25:42.5543276Z W0907 08:25:42.553452 82806 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] torch._dynamo.config.capture_scalar_outputs = True 2025-09-07T08:25:42.5543813Z W0907 08:25:42.553452 82806 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] or: 2025-09-07T08:25:42.5544430Z W0907 08:25:42.553452 82806 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1 2025-09-07T08:25:42.5545006Z W0907 08:25:42.553452 82806 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] to include these operations in the captured graph. 2025-09-07T08:25:42.5545490Z W0907 08:25:42.553452 82806 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T08:25:42.5545955Z W0907 08:25:42.553452 82806 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] Graph break: from user code at: 2025-09-07T08:25:42.5546795Z W0907 08:25:42.553452 82806 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2132, in torch_dynamo_resume_in_forward_at_2104 2025-09-07T08:25:42.5547669Z W0907 08:25:42.553452 82806 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] embedding_output = self.embeddings( 2025-09-07T08:25:42.5548416Z W0907 08:25:42.553452 82806 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 369, in forward 2025-09-07T08:25:42.5549250Z W0907 08:25:42.553452 82806 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] position_embeddings = self.position_embeddings(position_ids) 2025-09-07T08:25:42.5550075Z W0907 08:25:42.553452 82806 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 293, in forward 2025-09-07T08:25:42.5550847Z W0907 08:25:42.553452 82806 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] max_position_id = position_ids.max().item() 2025-09-07T08:25:42.5551315Z W0907 08:25:42.553452 82806 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T08:25:42.5551677Z W0907 08:25:42.553452 82806 site-packages/torch/_dynamo/variables/tensor.py:1048] [4/0] 2025-09-07T08:26:00.6097144Z pass 2025-09-07T08:26:00.6097556Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:03.2110060Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:26:03.2113594Z import pynvml # type: ignore[import] 2025-09-07T08:26:05.2210022Z 2025-09-07T08:26:10.9653827Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:26:10.9655356Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:26:11.0047545Z cpu eval hf_Roberta_base 2025-09-07T08:26:12.9226074Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:13.4540958Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:13.9850728Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:34.1059223Z pass 2025-09-07T08:26:34.1059641Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:37.7535363Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:26:37.7536643Z import pynvml # type: ignore[import] 2025-09-07T08:26:39.7797018Z 2025-09-07T08:26:42.6228388Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:26:42.6228673Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:26:42.6438561Z cpu eval hf_T5 2025-09-07T08:26:43.6408137Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:44.0964723Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:26:44.5426152Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:27:10.0498905Z pass 2025-09-07T08:27:10.0499288Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:27:12.9431824Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:27:12.9432690Z import pynvml # type: ignore[import] 2025-09-07T08:27:14.9568074Z 2025-09-07T08:27:24.5063370Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:27:24.5063653Z loading model: 0it [00:09, ?it/s] 2025-09-07T08:27:24.5469624Z cpu eval hf_T5_base 2025-09-07T08:27:39.0921829Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:27:45.7528039Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:27:52.2914291Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:28:55.2393670Z pass 2025-09-07T08:28:55.2394085Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:29:01.9380918Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:29:01.9382055Z import pynvml # type: ignore[import] 2025-09-07T08:29:03.9647178Z 2025-09-07T08:29:16.4534692Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:29:16.4535435Z loading model: 0it [00:12, ?it/s] 2025-09-07T08:29:16.5360029Z cpu eval hf_T5_large 2025-09-07T08:29:16.9380754Z pass_due_to_skip 2025-09-07T08:29:16.9383985Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:29:19.2481891Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:29:19.2483073Z import pynvml # type: ignore[import] 2025-09-07T08:29:21.2658390Z 2025-09-07T08:29:27.4875408Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:29:27.4875700Z loading model: 0it [00:06, ?it/s] 2025-09-07T08:29:27.5248830Z cpu eval hf_distil_whisper 2025-09-07T08:29:34.8381690Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:29:36.9567906Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:29:39.0777093Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:15.0977429Z pass 2025-09-07T08:30:15.0977888Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:18.3866224Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:30:18.3867132Z import pynvml # type: ignore[import] 2025-09-07T08:30:20.3856909Z 2025-09-07T08:30:20.5341989Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:30:20.5342275Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:30:20.5346795Z cpu eval lennard_jones 2025-09-07T08:30:20.5355920Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:20.5381722Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:20.5396247Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:27.6310332Z pass 2025-09-07T08:30:27.6312414Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:29.1302107Z accuracy pass_rate=91.30% 2025-09-07T08:30:29.1307305Z calls_captured gmean=0.00x mean=411.261x 2025-09-07T08:30:29.1310316Z unique_graphs gmean=0.00x mean=2.957x 2025-09-07T08:30:29.1312615Z graph_breaks gmean=0.00x mean=2.174x 2025-09-07T08:30:29.1314701Z unique_graph_breaks gmean=0.00x mean=0.565x 2025-09-07T08:30:29.1316800Z autograd_captures gmean=0.00x mean=0.000x 2025-09-07T08:30:29.1318825Z autograd_compiles gmean=0.00x mean=0.000x 2025-09-07T08:30:29.1320879Z cudagraph_skips gmean=0.00x mean=0.000x 2025-09-07T08:30:29.1321691Z compilation_latency mean=22.077 seconds 2025-09-07T08:30:29.6622584Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *freezing_cudagraphs-true* ]] 2025-09-07T08:30:29.6623406Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *freeze_autotune_cudagraphs-true* ]] 2025-09-07T08:30:29.6624171Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *aotinductor-true* ]] 2025-09-07T08:30:29.6624629Z + [[ inference == \i\n\f\e\r\e\n\c\e ]] 2025-09-07T08:30:29.6624837Z + [[ accuracy == \a\c\c\u\r\a\c\y ]] 2025-09-07T08:30:29.6625745Z + taskset -c 0-94 python benchmarks/dynamo/torchbench.py --accuracy --no-translation-validation --inference --bfloat16 --export --disable-cudagraphs --device cpu --total-partitions 4 --partition-id 1 --output /var/lib/jenkins/workspace/test/test-reports/inductor_export_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.csv 2025-09-07T08:30:30.0572363Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:30:30.0573570Z import pynvml # type: ignore[import] 2025-09-07T08:30:32.4980883Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:30:32.4981767Z import pynvml # type: ignore[import] 2025-09-07T08:30:34.5124074Z 2025-09-07T08:30:39.8422003Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:30:39.8422391Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:30:39.8701492Z cpu eval dlrm 2025-09-07T08:30:39.9976309Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:40.0225500Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:40.0483724Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:45.5746905Z pass 2025-09-07T08:30:45.5748884Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:30:46.9542872Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:30:46.9544272Z import pynvml # type: ignore[import] 2025-09-07T08:30:48.9749226Z 2025-09-07T08:30:51.6794760Z loading model: 0it [00:00, ?it/s]WARNING:common:Model doctr_det_predictor does not support bfloat16, running with amp instead 2025-09-07T08:30:52.3054419Z 2025-09-07T08:30:52.3059413Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:30:52.3059838Z WARNING:common:Model doctr_det_predictor does not support bfloat16, running with amp instead 2025-09-07T08:30:52.3060183Z cpu eval doctr_det_predictor 2025-09-07T08:31:06.8494342Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:06.8494874Z WARNING:common:Model doctr_det_predictor does not support bfloat16, running with amp instead 2025-09-07T08:31:07.7525158Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:08.6492414Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:09.7771272Z ERROR:common: 2025-09-07T08:31:09.7771515Z Traceback (most recent call last): 2025-09-07T08:31:09.7771934Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 2320, in check_accuracy 2025-09-07T08:31:09.7772300Z optimized_model_iter_fn = optimize_ctx( 2025-09-07T08:31:09.7772619Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 1523, in export 2025-09-07T08:31:09.7772941Z ep = torch.export.export( 2025-09-07T08:31:09.7773289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 311, in export 2025-09-07T08:31:09.7773627Z raise e 2025-09-07T08:31:09.7773903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 277, in export 2025-09-07T08:31:09.7774239Z return _export( 2025-09-07T08:31:09.7774528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:31:09.7774843Z raise e 2025-09-07T08:31:09.7775112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:31:09.7775424Z ep = fn(*args, **kwargs) 2025-09-07T08:31:09.7776160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:31:09.7776519Z return fn(*args, **kwargs) 2025-09-07T08:31:09.7776833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2255, in _export 2025-09-07T08:31:09.7777173Z ep = _export_for_training( 2025-09-07T08:31:09.7777666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:31:09.7777984Z raise e 2025-09-07T08:31:09.7778256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:31:09.7778569Z ep = fn(*args, **kwargs) 2025-09-07T08:31:09.7778899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:31:09.7779248Z return fn(*args, **kwargs) 2025-09-07T08:31:09.7779591Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2071, in _export_for_training 2025-09-07T08:31:09.7779956Z export_artifact = export_func( 2025-09-07T08:31:09.7780287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1415, in _strict_export 2025-09-07T08:31:09.7780643Z gm_torch_level = _export_to_torch_ir( 2025-09-07T08:31:09.7781092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 812, in _export_to_torch_ir 2025-09-07T08:31:09.7781469Z gm_torch_level, _ = torch._dynamo.export( 2025-09-07T08:31:09.7781836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 2002, in inner 2025-09-07T08:31:09.7782181Z result_traced = opt_f(*args, **kwargs) 2025-09-07T08:31:09.7782590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 414, in __call__ 2025-09-07T08:31:09.7782943Z return super().__call__(*args, **kwargs) 2025-09-07T08:31:09.7783327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl 2025-09-07T08:31:09.7783699Z return self._call_impl(*args, **kwargs) 2025-09-07T08:31:09.7784046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:31:09.7784398Z return forward_call(*args, **kwargs) 2025-09-07T08:31:09.7784756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 841, in compile_wrapper 2025-09-07T08:31:09.7785169Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-09-07T08:31:09.7785498Z torch._dynamo.exc.Unsupported: Attempted to call function marked as skipped 2025-09-07T08:31:09.7786177Z Explanation: Dynamo does not know how to trace the builtin `.morphologyEx.` This function is either a Python builtin (e.g. _warnings.warn) or a third-party C/C++ Python extension (perhaps created with pybind). 2025-09-07T08:31:09.7786987Z Hint: If it is a Python builtin, please file an issue on GitHub so the PyTorch team can add support for it and see the next case for a workaround. 2025-09-07T08:31:09.7787902Z Hint: If it is a third-party C/C++ Python extension, please either wrap it into a PyTorch-understood custom operator (see https://pytorch.org/tutorials/advanced/custom_ops_landing_page.html for more details) or, if it is traceable, use `torch.compiler.allow_in_graph`. 2025-09-07T08:31:09.7788551Z 2025-09-07T08:31:09.7788762Z Developer debug context: module: , qualname: morphologyEx, skip reason: 2025-09-07T08:31:09.7789034Z 2025-09-07T08:31:09.7789317Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0007.html 2025-09-07T08:31:09.7789635Z 2025-09-07T08:31:09.7789701Z from user code: 2025-09-07T08:31:09.7790119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 211, in forward 2025-09-07T08:31:09.7790710Z for preds in self.postprocessor(prob_map.detach().cpu().permute((0, 2, 3, 1)).numpy()) 2025-09-07T08:31:09.7791166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/core.py", line 90, in __call__ 2025-09-07T08:31:09.7791507Z bin_map = [ 2025-09-07T08:31:09.7792563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/core.py", line 91, in 2025-09-07T08:31:09.7792899Z [ 2025-09-07T08:31:09.7793222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/core.py", line 92, in 2025-09-07T08:31:09.7793655Z cv2.morphologyEx(bmap[..., idx], cv2.MORPH_OPEN, self._opening_kernel) 2025-09-07T08:31:09.7793851Z 2025-09-07T08:31:09.7794212Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-09-07T08:31:09.7794614Z 2025-09-07T08:31:09.7794760Z TorchDynamo optimized model failed to run because of following error 2025-09-07T08:31:10.0514921Z fail_to_run 2025-09-07T08:31:10.0515277Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:11.9805696Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:31:11.9806589Z import pynvml # type: ignore[import] 2025-09-07T08:31:14.0026114Z 2025-09-07T08:31:16.6193757Z loading model: 0it [00:00, ?it/s]WARNING:common:Model doctr_reco_predictor does not support bfloat16, running with amp instead 2025-09-07T08:31:16.6624035Z 2025-09-07T08:31:16.6624310Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:31:16.6624710Z WARNING:common:Model doctr_reco_predictor does not support bfloat16, running with amp instead 2025-09-07T08:31:16.6625111Z cpu eval doctr_reco_predictor 2025-09-07T08:31:17.1446417Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:17.1447265Z WARNING:common:Model doctr_reco_predictor does not support bfloat16, running with amp instead 2025-09-07T08:31:17.2054399Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:17.2529831Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:17.8048293Z ERROR:common: 2025-09-07T08:31:17.8048528Z Traceback (most recent call last): 2025-09-07T08:31:17.8049010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/variables/tensor.py", line 1435, in evaluate_expr 2025-09-07T08:31:17.8049428Z return guard_scalar(self.sym_num) 2025-09-07T08:31:17.8049839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/symbolic_shapes.py", line 1535, in guard_scalar 2025-09-07T08:31:17.8050255Z return guard_int(a) 2025-09-07T08:31:17.8050627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/symbolic_shapes.py", line 1732, in guard_int 2025-09-07T08:31:17.8051085Z return a.node.guard_int("", 0) # NB: uses Python backtrace 2025-09-07T08:31:17.8051546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/sym_node.py", line 518, in guard_int 2025-09-07T08:31:17.8051907Z r = self.evaluate() 2025-09-07T08:31:17.8052249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/sym_node.py", line 512, in evaluate 2025-09-07T08:31:17.8052706Z return self.shape_env.evaluate_sym_node(self, size_oblivious) 2025-09-07T08:31:17.8053178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/symbolic_shapes.py", line 7239, in evaluate_sym_node 2025-09-07T08:31:17.8053928Z return self.evaluate_expr( 2025-09-07T08:31:17.8054310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/symbolic_shapes.py", line 7339, in evaluate_expr 2025-09-07T08:31:17.8054713Z return self._inner_evaluate_expr( 2025-09-07T08:31:17.8055083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/recording.py", line 272, in wrapper 2025-09-07T08:31:17.8055546Z return retlog(fn(*args, **kwargs)) 2025-09-07T08:31:17.8055951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/symbolic_shapes.py", line 7362, in _inner_evaluate_expr 2025-09-07T08:31:17.8056362Z return self._evaluate_expr( 2025-09-07T08:31:17.8056751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/fx/experimental/symbolic_shapes.py", line 7580, in _evaluate_expr 2025-09-07T08:31:17.8057158Z raise self._make_data_dependent_error( 2025-09-07T08:31:17.8057779Z torch.fx.experimental.symbolic_shapes.GuardOnDataDependentSymNode: Could not extract specialized integer from data-dependent expression u0 (unhinted: u0). (Size-like symbols: none) 2025-09-07T08:31:17.8058260Z 2025-09-07T08:31:17.8058649Z Caused by: decode_sequence([k for k, _ in groupby(seq.tolist()) if k != blank], vocab) # doctr/models/recognition/crnn/pytorch.py:79 in (_dynamo/variables/tensor.py:1435 in evaluate_expr) 2025-09-07T08:31:17.8059258Z For more information, run with TORCH_LOGS="dynamic" 2025-09-07T08:31:17.8059607Z For extended logs when we create symbols, also add TORCHDYNAMO_EXTENDED_DEBUG_CREATE_SYMBOL="u0" 2025-09-07T08:31:17.8060015Z If you suspect the guard was triggered from C++, add TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:31:17.8060611Z For more debugging help, see https://docs.google.com/document/d/1HSuTTVvYH1pTew89Rtpeu84Ht3nQEFTYhAX3Ypa_xJs/edit?usp=sharing 2025-09-07T08:31:17.8060955Z 2025-09-07T08:31:17.8061038Z User Stack (most recent call last): 2025-09-07T08:31:17.8061240Z (snipped, see stack below for prefix) 2025-09-07T08:31:17.8061641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 224, in forward 2025-09-07T08:31:17.8062048Z out["preds"] = self.postprocessor(logits) 2025-09-07T08:31:17.8062439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 99, in __call__ 2025-09-07T08:31:17.8062905Z return self.ctc_best_path(logits=logits, vocab=self.vocab, blank=len(self.vocab)) 2025-09-07T08:31:17.8063391Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 78, in ctc_best_path 2025-09-07T08:31:17.8063773Z words = [ 2025-09-07T08:31:17.8064113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 79, in 2025-09-07T08:31:17.8064562Z decode_sequence([k for k, _ in groupby(seq.tolist()) if k != blank], vocab) 2025-09-07T08:31:17.8064755Z 2025-09-07T08:31:17.8064873Z For C++ stack trace, run with TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:31:17.8065053Z 2025-09-07T08:31:17.8065184Z During handling of the above exception, another exception occurred: 2025-09-07T08:31:17.8065378Z 2025-09-07T08:31:17.8065792Z torch._dynamo.exc.UserError: Consider annotating your code using torch._check*(). Could not extract specialized integer from data-dependent expression u0 (unhinted: u0). (Size-like symbols: none) 2025-09-07T08:31:17.8066264Z 2025-09-07T08:31:17.8066641Z Caused by: decode_sequence([k for k, _ in groupby(seq.tolist()) if k != blank], vocab) # doctr/models/recognition/crnn/pytorch.py:79 in (_dynamo/variables/tensor.py:1435 in evaluate_expr) 2025-09-07T08:31:17.8067163Z For more information, run with TORCH_LOGS="dynamic" 2025-09-07T08:31:17.8067503Z For extended logs when we create symbols, also add TORCHDYNAMO_EXTENDED_DEBUG_CREATE_SYMBOL="u0" 2025-09-07T08:31:17.8067900Z If you suspect the guard was triggered from C++, add TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:31:17.8068468Z For more debugging help, see https://docs.google.com/document/d/1HSuTTVvYH1pTew89Rtpeu84Ht3nQEFTYhAX3Ypa_xJs/edit?usp=sharing 2025-09-07T08:31:17.8068812Z 2025-09-07T08:31:17.8068890Z User Stack (most recent call last): 2025-09-07T08:31:17.8069097Z (snipped, see stack below for prefix) 2025-09-07T08:31:17.8069495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 224, in forward 2025-09-07T08:31:17.8069935Z out["preds"] = self.postprocessor(logits) 2025-09-07T08:31:17.8070319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 99, in __call__ 2025-09-07T08:31:17.8070788Z return self.ctc_best_path(logits=logits, vocab=self.vocab, blank=len(self.vocab)) 2025-09-07T08:31:17.8071278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 78, in ctc_best_path 2025-09-07T08:31:17.8071659Z words = [ 2025-09-07T08:31:17.8072003Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 79, in 2025-09-07T08:31:17.8072442Z decode_sequence([k for k, _ in groupby(seq.tolist()) if k != blank], vocab) 2025-09-07T08:31:17.8072641Z 2025-09-07T08:31:17.8072756Z For C++ stack trace, run with TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:31:17.8073248Z For more information about this error, see: https://pytorch.org/docs/main/generated/exportdb/index.html#constrain-as-size-example 2025-09-07T08:31:17.8073585Z 2025-09-07T08:31:17.8073731Z The above exception was the direct cause of the following exception: 2025-09-07T08:31:17.8073920Z 2025-09-07T08:31:17.8074003Z Traceback (most recent call last): 2025-09-07T08:31:17.8074352Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 2320, in check_accuracy 2025-09-07T08:31:17.8074693Z optimized_model_iter_fn = optimize_ctx( 2025-09-07T08:31:17.8075004Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 1523, in export 2025-09-07T08:31:17.8075313Z ep = torch.export.export( 2025-09-07T08:31:17.8075636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 311, in export 2025-09-07T08:31:17.8075951Z raise e 2025-09-07T08:31:17.8076224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 277, in export 2025-09-07T08:31:17.8076543Z return _export( 2025-09-07T08:31:17.8076828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:31:17.8077132Z raise e 2025-09-07T08:31:17.8077397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:31:17.8077721Z ep = fn(*args, **kwargs) 2025-09-07T08:31:17.8078062Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:31:17.8078409Z return fn(*args, **kwargs) 2025-09-07T08:31:17.8078727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2255, in _export 2025-09-07T08:31:17.8079048Z ep = _export_for_training( 2025-09-07T08:31:17.8079348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:31:17.8079656Z raise e 2025-09-07T08:31:17.8079918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:31:17.8080232Z ep = fn(*args, **kwargs) 2025-09-07T08:31:17.8080558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:31:17.8080902Z return fn(*args, **kwargs) 2025-09-07T08:31:17.8081235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2071, in _export_for_training 2025-09-07T08:31:17.8081601Z export_artifact = export_func( 2025-09-07T08:31:17.8081986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1415, in _strict_export 2025-09-07T08:31:17.8082344Z gm_torch_level = _export_to_torch_ir( 2025-09-07T08:31:17.8082705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 812, in _export_to_torch_ir 2025-09-07T08:31:17.8083110Z gm_torch_level, _ = torch._dynamo.export( 2025-09-07T08:31:17.8083461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 2002, in inner 2025-09-07T08:31:17.8083800Z result_traced = opt_f(*args, **kwargs) 2025-09-07T08:31:17.8084145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 414, in __call__ 2025-09-07T08:31:17.8084492Z return super().__call__(*args, **kwargs) 2025-09-07T08:31:17.8084874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl 2025-09-07T08:31:17.8085249Z return self._call_impl(*args, **kwargs) 2025-09-07T08:31:17.8085599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:31:17.8085952Z return forward_call(*args, **kwargs) 2025-09-07T08:31:17.8086305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 841, in compile_wrapper 2025-09-07T08:31:17.8086755Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-09-07T08:31:17.8087125Z torch._dynamo.exc.Unsupported: Unexpected failure during itertools.groupby() iteration 2025-09-07T08:31:17.8087495Z Explanation: Unexpected failure in invoking function during groupby 2025-09-07T08:31:17.8088073Z Hint: It may be possible to write Dynamo tracing rules for this code. Please report an issue to PyTorch if you encounter this graph break often and it is causing performance issues. 2025-09-07T08:31:17.8088482Z 2025-09-07T08:31:17.8088723Z Developer debug context: call_function ItertoolsVariable() [ListVariable(length=32)] {} 2025-09-07T08:31:17.8089030Z 2025-09-07T08:31:17.8089293Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0115.html 2025-09-07T08:31:17.8089619Z 2025-09-07T08:31:17.8089680Z from user code: 2025-09-07T08:31:17.8090039Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 224, in forward 2025-09-07T08:31:17.8090441Z out["preds"] = self.postprocessor(logits) 2025-09-07T08:31:17.8090826Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 99, in __call__ 2025-09-07T08:31:17.8091298Z return self.ctc_best_path(logits=logits, vocab=self.vocab, blank=len(self.vocab)) 2025-09-07T08:31:17.8091775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 78, in ctc_best_path 2025-09-07T08:31:17.8092158Z words = [ 2025-09-07T08:31:17.8092497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/recognition/crnn/pytorch.py", line 79, in 2025-09-07T08:31:17.8092945Z decode_sequence([k for k, _ in groupby(seq.tolist()) if k != blank], vocab) 2025-09-07T08:31:17.8093143Z 2025-09-07T08:31:17.8093492Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-09-07T08:31:17.8093898Z 2025-09-07T08:31:17.8094034Z TorchDynamo optimized model failed to run because of following error 2025-09-07T08:31:17.8704525Z fail_to_run 2025-09-07T08:31:17.8705301Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:19.1807157Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:31:19.1808335Z import pynvml # type: ignore[import] 2025-09-07T08:31:21.1906229Z 2025-09-07T08:31:21.2427920Z loading model: 0it [00:00, ?it/s]Gym has been unmaintained since 2022 and does not support NumPy 2.0 amongst other critical functionality. 2025-09-07T08:31:21.2428948Z Please upgrade to Gymnasium, the maintained drop-in replacement of Gym, or contact the authors of your software and request that they upgrade. 2025-09-07T08:31:21.2429553Z Users of this version of Gym should be able to simply replace 'import gym' with 'import gymnasium as gym' in the vast majority of cases. 2025-09-07T08:31:21.2430159Z See the migration guide at https://gymnasium.farama.org/introduction/migration_guide/ for additional information. 2025-09-07T08:31:22.1373929Z 2025-09-07T08:31:22.1374341Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:31:22.1389027Z cpu eval drq 2025-09-07T08:31:22.1565645Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:22.1674720Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:22.1752793Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:27.0083198Z pass 2025-09-07T08:31:27.0083817Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:28.2223612Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:31:28.2224506Z import pynvml # type: ignore[import] 2025-09-07T08:31:30.2383836Z 2025-09-07T08:31:31.7919237Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:31:31.7919575Z loading model: 0it [00:01, ?it/s] 2025-09-07T08:31:31.8097714Z cpu eval fastNLP_Bert 2025-09-07T08:31:32.8166356Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:33.2079154Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:33.5911233Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:33.9041835Z ERROR:common: 2025-09-07T08:31:33.9042063Z Traceback (most recent call last): 2025-09-07T08:31:33.9042431Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 2320, in check_accuracy 2025-09-07T08:31:33.9042803Z optimized_model_iter_fn = optimize_ctx( 2025-09-07T08:31:33.9043144Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 1523, in export 2025-09-07T08:31:33.9043465Z ep = torch.export.export( 2025-09-07T08:31:33.9043827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 311, in export 2025-09-07T08:31:33.9044177Z raise e 2025-09-07T08:31:33.9044470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 277, in export 2025-09-07T08:31:33.9044784Z return _export( 2025-09-07T08:31:33.9045093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:31:33.9045412Z raise e 2025-09-07T08:31:33.9045683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:31:33.9046023Z ep = fn(*args, **kwargs) 2025-09-07T08:31:33.9046376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:31:33.9046740Z return fn(*args, **kwargs) 2025-09-07T08:31:33.9047045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2255, in _export 2025-09-07T08:31:33.9047667Z ep = _export_for_training( 2025-09-07T08:31:33.9047966Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:31:33.9048276Z raise e 2025-09-07T08:31:33.9048545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:31:33.9048989Z ep = fn(*args, **kwargs) 2025-09-07T08:31:33.9049310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:31:33.9049678Z return fn(*args, **kwargs) 2025-09-07T08:31:33.9050037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2071, in _export_for_training 2025-09-07T08:31:33.9050408Z export_artifact = export_func( 2025-09-07T08:31:33.9050745Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1415, in _strict_export 2025-09-07T08:31:33.9051099Z gm_torch_level = _export_to_torch_ir( 2025-09-07T08:31:33.9051455Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 812, in _export_to_torch_ir 2025-09-07T08:31:33.9051838Z gm_torch_level, _ = torch._dynamo.export( 2025-09-07T08:31:33.9052188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 2002, in inner 2025-09-07T08:31:33.9052606Z result_traced = opt_f(*args, **kwargs) 2025-09-07T08:31:33.9052941Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 414, in __call__ 2025-09-07T08:31:33.9053288Z return super().__call__(*args, **kwargs) 2025-09-07T08:31:33.9053728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl 2025-09-07T08:31:33.9054109Z return self._call_impl(*args, **kwargs) 2025-09-07T08:31:33.9054446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:31:33.9054797Z return forward_call(*args, **kwargs) 2025-09-07T08:31:33.9055154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 841, in compile_wrapper 2025-09-07T08:31:33.9055570Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-09-07T08:31:33.9056146Z torch._dynamo.exc.UserError: Consider annotating your code using torch._check*(). Could not guard on data-dependent expression u0 + 2 > 512 (unhinted: u0 + 2 > 512). (Size-like symbols: none) 2025-09-07T08:31:33.9056574Z 2025-09-07T08:31:33.9057195Z consider using data-dependent friendly APIs such as guard_or_false, guard_or_true and statically_known_trueCaused by: if max_word_piece_length + 2 > self._max_position_embeddings: # fastNLP/embeddings/bert_embedding.py:446 in forward (_dynamo/variables/tensor.py:1435 in evaluate_expr) 2025-09-07T08:31:33.9057996Z For more information, run with TORCH_LOGS="dynamic" 2025-09-07T08:31:33.9058340Z For extended logs when we create symbols, also add TORCHDYNAMO_EXTENDED_DEBUG_CREATE_SYMBOL="u0" 2025-09-07T08:31:33.9058748Z If you suspect the guard was triggered from C++, add TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:31:33.9059280Z For more debugging help, see https://docs.google.com/document/d/1HSuTTVvYH1pTew89Rtpeu84Ht3nQEFTYhAX3Ypa_xJs/edit?usp=sharing 2025-09-07T08:31:33.9059624Z 2025-09-07T08:31:33.9059704Z User Stack (most recent call last): 2025-09-07T08:31:33.9059903Z (snipped, see stack below for prefix) 2025-09-07T08:31:33.9060249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/models/bert.py", line 265, in forward 2025-09-07T08:31:33.9060584Z sequence_output = self.bert(words) 2025-09-07T08:31:33.9060939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:31:33.9061290Z return forward_call(*args, **kwargs) 2025-09-07T08:31:33.9061662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 137, in forward 2025-09-07T08:31:33.9062076Z outputs = self.model(words) 2025-09-07T08:31:33.9062408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:31:33.9062750Z return forward_call(*args, **kwargs) 2025-09-07T08:31:33.9063140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 446, in forward 2025-09-07T08:31:33.9063559Z if max_word_piece_length + 2 > self._max_position_embeddings: 2025-09-07T08:31:33.9063732Z 2025-09-07T08:31:33.9063845Z For C++ stack trace, run with TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:31:33.9064303Z For more information about this error, see: https://pytorch.org/docs/main/generated/exportdb/index.html#constrain-as-size-example 2025-09-07T08:31:33.9064637Z 2025-09-07T08:31:33.9064701Z from user code: 2025-09-07T08:31:33.9064988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/models/bert.py", line 265, in forward 2025-09-07T08:31:33.9065327Z sequence_output = self.bert(words) 2025-09-07T08:31:33.9065669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:31:33.9066022Z return forward_call(*args, **kwargs) 2025-09-07T08:31:33.9066488Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 137, in forward 2025-09-07T08:31:33.9066850Z outputs = self.model(words) 2025-09-07T08:31:33.9067175Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:31:33.9067523Z return forward_call(*args, **kwargs) 2025-09-07T08:31:33.9067913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/fastNLP/embeddings/bert_embedding.py", line 446, in forward 2025-09-07T08:31:33.9068319Z if max_word_piece_length + 2 > self._max_position_embeddings: 2025-09-07T08:31:33.9068500Z 2025-09-07T08:31:33.9068853Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-09-07T08:31:33.9069265Z 2025-09-07T08:31:33.9069268Z 2025-09-07T08:31:33.9069766Z The error above occurred when calling torch.export.export. If you would like to view some more information about this error, and get a list of all other errors that may occur in your export call, you can replace your `export()` call with `draft_export()`. 2025-09-07T08:31:33.9070454Z TorchDynamo optimized model failed to run because of following error 2025-09-07T08:31:34.0035733Z fail_to_run 2025-09-07T08:31:34.0036083Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:35.3192108Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:31:35.3192978Z import pynvml # type: ignore[import] 2025-09-07T08:31:37.3286175Z 2025-09-07T08:31:37.7620917Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:31:37.7621227Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:31:37.7675431Z cpu eval functorch_dp_cifar10 2025-09-07T08:31:37.8337248Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:37.8613411Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:37.8835118Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:42.8462477Z pass 2025-09-07T08:31:42.8462881Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:44.0127653Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:31:44.0129038Z import pynvml # type: ignore[import] 2025-09-07T08:31:46.0198043Z 2025-09-07T08:31:46.2146266Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:31:46.2146549Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:31:46.2151254Z cpu eval functorch_maml_omniglot 2025-09-07T08:31:46.2248544Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:46.2329143Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:46.2376390Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:50.9351517Z pass 2025-09-07T08:31:50.9351939Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:52.0606863Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:31:52.0608178Z import pynvml # type: ignore[import] 2025-09-07T08:31:54.0655443Z 2025-09-07T08:31:55.7439131Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:31:55.7439406Z loading model: 0it [00:01, ?it/s] 2025-09-07T08:31:55.7487924Z cpu eval hf_Albert 2025-09-07T08:31:56.6682751Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:56.9089498Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:31:57.1361737Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:03.7538543Z pass 2025-09-07T08:32:03.7538938Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:05.2152065Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:32:05.2153188Z import pynvml # type: ignore[import] 2025-09-07T08:32:07.2278241Z 2025-09-07T08:32:11.3968518Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:32:11.3969105Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:32:11.4292562Z cpu eval hf_Bart 2025-09-07T08:32:12.7757879Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:13.1312428Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:13.4788402Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:20.5306344Z pass 2025-09-07T08:32:20.5306736Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:22.2254171Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:32:22.2255056Z import pynvml # type: ignore[import] 2025-09-07T08:32:24.2291516Z 2025-09-07T08:32:27.3692408Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:32:27.3692691Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:32:27.3925683Z cpu eval hf_Bert 2025-09-07T08:32:28.4271207Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:28.7161335Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:28.9876461Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:35.4566143Z pass 2025-09-07T08:32:35.4566603Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:37.0854851Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:32:37.0855726Z import pynvml # type: ignore[import] 2025-09-07T08:32:39.1048795Z 2025-09-07T08:32:44.4912177Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:32:44.4912540Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:32:44.5308931Z cpu eval hf_Bert_large 2025-09-07T08:32:46.7955258Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:47.4461831Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:48.0870442Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:56.5789338Z pass 2025-09-07T08:32:56.5789729Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:32:58.3530738Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:32:58.3531642Z import pynvml # type: ignore[import] 2025-09-07T08:33:00.3749219Z 2025-09-07T08:33:01.7805413Z loading model: 0it [00:00, ?it/s]BigBirdForMaskedLM has generative capabilities, as `prepare_inputs_for_generation` is explicitly defined. However, it doesn't directly inherit from `GenerationMixin`. From 👉v4.50👈 onwards, `PreTrainedModel` will NOT inherit from `GenerationMixin`, and this model will lose the ability to call `generate` and other related functions. 2025-09-07T08:33:01.7806725Z - If you're using `trust_remote_code=True`, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes 2025-09-07T08:33:01.7807561Z - If you are the owner of the model architecture code, please modify your model class such that it inherits from `GenerationMixin` (after `PreTrainedModel`, otherwise you'll get an exception). 2025-09-07T08:33:01.7808213Z - If you are not the owner of the model architecture class, please contact the model code owner to update it. 2025-09-07T08:33:01.7809509Z WARNING:transformers.modeling_utils:BigBirdForMaskedLM has generative capabilities, as `prepare_inputs_for_generation` is explicitly defined. However, it doesn't directly inherit from `GenerationMixin`. From 👉v4.50👈 onwards, `PreTrainedModel` will NOT inherit from `GenerationMixin`, and this model will lose the ability to call `generate` and other related functions. 2025-09-07T08:33:01.7810732Z - If you're using `trust_remote_code=True`, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes 2025-09-07T08:33:01.7811547Z - If you are the owner of the model architecture code, please modify your model class such that it inherits from `GenerationMixin` (after `PreTrainedModel`, otherwise you'll get an exception). 2025-09-07T08:33:01.7812181Z - If you are not the owner of the model architecture class, please contact the model code owner to update it. 2025-09-07T08:33:03.6655362Z Input ids are automatically padded from 819 to 832 to be a multiple of `config.block_size`: 64 2025-09-07T08:33:03.6656096Z WARNING:transformers.models.big_bird.modeling_big_bird:Input ids are automatically padded from 819 to 832 to be a multiple of `config.block_size`: 64 2025-09-07T08:33:05.5048273Z 2025-09-07T08:33:05.5049427Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:33:05.5310013Z cpu eval hf_BigBird 2025-09-07T08:33:08.1400032Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:11.2312016Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:14.2488311Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:16.1929308Z ERROR:common: 2025-09-07T08:33:16.1929547Z Traceback (most recent call last): 2025-09-07T08:33:16.1929927Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 2320, in check_accuracy 2025-09-07T08:33:16.1930311Z optimized_model_iter_fn = optimize_ctx( 2025-09-07T08:33:16.1930635Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 1523, in export 2025-09-07T08:33:16.1930943Z ep = torch.export.export( 2025-09-07T08:33:16.1931617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 311, in export 2025-09-07T08:33:16.1931968Z raise e 2025-09-07T08:33:16.1932258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 277, in export 2025-09-07T08:33:16.1932577Z return _export( 2025-09-07T08:33:16.1932871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:33:16.1933273Z raise e 2025-09-07T08:33:16.1933553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:33:16.1933872Z ep = fn(*args, **kwargs) 2025-09-07T08:33:16.1934216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:33:16.1934583Z return fn(*args, **kwargs) 2025-09-07T08:33:16.1934892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2255, in _export 2025-09-07T08:33:16.1935211Z ep = _export_for_training( 2025-09-07T08:33:16.1935503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:33:16.1935812Z raise e 2025-09-07T08:33:16.1936077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:33:16.1936394Z ep = fn(*args, **kwargs) 2025-09-07T08:33:16.1936711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:33:16.1937053Z return fn(*args, **kwargs) 2025-09-07T08:33:16.1937393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2071, in _export_for_training 2025-09-07T08:33:16.1937841Z export_artifact = export_func( 2025-09-07T08:33:16.1938180Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1415, in _strict_export 2025-09-07T08:33:16.1938530Z gm_torch_level = _export_to_torch_ir( 2025-09-07T08:33:16.1938892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 812, in _export_to_torch_ir 2025-09-07T08:33:16.1939260Z gm_torch_level, _ = torch._dynamo.export( 2025-09-07T08:33:16.1939610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 2002, in inner 2025-09-07T08:33:16.1939948Z result_traced = opt_f(*args, **kwargs) 2025-09-07T08:33:16.1940289Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 414, in __call__ 2025-09-07T08:33:16.1940634Z return super().__call__(*args, **kwargs) 2025-09-07T08:33:16.1941112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl 2025-09-07T08:33:16.1941492Z return self._call_impl(*args, **kwargs) 2025-09-07T08:33:16.1941830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:33:16.1942286Z return forward_call(*args, **kwargs) 2025-09-07T08:33:16.1942640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 841, in compile_wrapper 2025-09-07T08:33:16.1943057Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-09-07T08:33:16.1943519Z torch._dynamo.exc.UserError: Could not guard on data-dependent expression Eq(u1, u0) (unhinted: Eq(u1, u0)). (Size-like symbols: u0, u1) 2025-09-07T08:33:16.1943848Z 2025-09-07T08:33:16.1944465Z consider using data-dependent friendly APIs such as guard_or_false, guard_or_true and statically_known_trueCaused by: rand_attn = np.stack(rand_attn, axis=0) # transformers/models/big_bird/modeling_big_bird.py:574 in bigbird_block_sparse_attention (_refs/__init__.py:3992 in _check_stack_inputs) 2025-09-07T08:33:16.1945237Z For more information, run with TORCH_LOGS="dynamic" 2025-09-07T08:33:16.1945587Z For extended logs when we create symbols, also add TORCHDYNAMO_EXTENDED_DEBUG_CREATE_SYMBOL="u0,u1" 2025-09-07T08:33:16.1946044Z If you suspect the guard was triggered from C++, add TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:33:16.1946587Z For more debugging help, see https://docs.google.com/document/d/1HSuTTVvYH1pTew89Rtpeu84Ht3nQEFTYhAX3Ypa_xJs/edit?usp=sharing 2025-09-07T08:33:16.1946950Z 2025-09-07T08:33:16.1947031Z User Stack (most recent call last): 2025-09-07T08:33:16.1947238Z (snipped, see stack below for prefix) 2025-09-07T08:33:16.1947697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2303, in forward 2025-09-07T08:33:16.1948110Z outputs = self.bert( 2025-09-07T08:33:16.1948443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:33:16.1948796Z return forward_call(*args, **kwargs) 2025-09-07T08:33:16.1949217Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1991, in forward 2025-09-07T08:33:16.1949623Z encoder_outputs = self.encoder( 2025-09-07T08:33:16.1949958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:33:16.1950301Z return forward_call(*args, **kwargs) 2025-09-07T08:33:16.1950680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1585, in forward 2025-09-07T08:33:16.1951078Z layer_outputs = layer_module( 2025-09-07T08:33:16.1951410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:33:16.1951752Z return forward_call(*args, **kwargs) 2025-09-07T08:33:16.1952134Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1468, in forward 2025-09-07T08:33:16.1952534Z self_attention_outputs = self.attention( 2025-09-07T08:33:16.1952882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:33:16.1953227Z return forward_call(*args, **kwargs) 2025-09-07T08:33:16.1953612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1381, in forward 2025-09-07T08:33:16.1954004Z self_outputs = self.self( 2025-09-07T08:33:16.1954326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:33:16.1954674Z return forward_call(*args, **kwargs) 2025-09-07T08:33:16.1955065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 455, in forward 2025-09-07T08:33:16.1955576Z context_layer, attention_probs = self.bigbird_block_sparse_attention( 2025-09-07T08:33:16.1956095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 574, in bigbird_block_sparse_attention 2025-09-07T08:33:16.1956591Z rand_attn = np.stack(rand_attn, axis=0) 2025-09-07T08:33:16.1956729Z 2025-09-07T08:33:16.1956846Z For C++ stack trace, run with TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:33:16.1957305Z For more information about this error, see: https://pytorch.org/docs/main/generated/exportdb/index.html#constrain-as-size-example 2025-09-07T08:33:16.1957634Z 2025-09-07T08:33:16.1957698Z from user code: 2025-09-07T08:33:16.1958069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2303, in forward 2025-09-07T08:33:16.1958465Z outputs = self.bert( 2025-09-07T08:33:16.1958789Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:33:16.1959143Z return forward_call(*args, **kwargs) 2025-09-07T08:33:16.1959534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1991, in forward 2025-09-07T08:33:16.1959960Z encoder_outputs = self.encoder( 2025-09-07T08:33:16.1960297Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:33:16.1960639Z return forward_call(*args, **kwargs) 2025-09-07T08:33:16.1961026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1585, in forward 2025-09-07T08:33:16.1961456Z layer_outputs = layer_module( 2025-09-07T08:33:16.1961781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:33:16.1962131Z return forward_call(*args, **kwargs) 2025-09-07T08:33:16.1962517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1468, in forward 2025-09-07T08:33:16.1962926Z self_attention_outputs = self.attention( 2025-09-07T08:33:16.1963267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:33:16.1963617Z return forward_call(*args, **kwargs) 2025-09-07T08:33:16.1964000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1381, in forward 2025-09-07T08:33:16.1964394Z self_outputs = self.self( 2025-09-07T08:33:16.1964718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:33:16.1965053Z return forward_call(*args, **kwargs) 2025-09-07T08:33:16.1965440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 455, in forward 2025-09-07T08:33:16.1965905Z context_layer, attention_probs = self.bigbird_block_sparse_attention( 2025-09-07T08:33:16.1966418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 574, in bigbird_block_sparse_attention 2025-09-07T08:33:16.1966876Z rand_attn = np.stack(rand_attn, axis=0) 2025-09-07T08:33:16.1967007Z 2025-09-07T08:33:16.1967362Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-09-07T08:33:16.1967771Z 2025-09-07T08:33:16.1967773Z 2025-09-07T08:33:16.1968271Z The error above occurred when calling torch.export.export. If you would like to view some more information about this error, and get a list of all other errors that may occur in your export call, you can replace your `export()` call with `draft_export()`. 2025-09-07T08:33:16.1969013Z TorchDynamo optimized model failed to run because of following error 2025-09-07T08:33:16.3898457Z fail_to_run 2025-09-07T08:33:16.3898803Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:18.1733207Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:33:18.1734436Z import pynvml # type: ignore[import] 2025-09-07T08:33:20.1779176Z 2025-09-07T08:33:22.7969366Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:33:22.7970024Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:33:22.8104910Z cpu eval hf_DistilBert 2025-09-07T08:33:23.3980680Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:23.5624760Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:23.7161624Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:29.0128842Z pass 2025-09-07T08:33:29.0129551Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:30.5699476Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:33:30.5700363Z import pynvml # type: ignore[import] 2025-09-07T08:33:32.5892259Z 2025-09-07T08:33:37.2496321Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:33:37.2496659Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:33:37.2674826Z cpu eval hf_GPT2 2025-09-07T08:33:37.8203232Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:37.9820095Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:38.1362671Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:45.6604116Z pass 2025-09-07T08:33:45.6604515Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:33:47.3320872Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:33:47.3321767Z import pynvml # type: ignore[import] 2025-09-07T08:33:49.3516252Z 2025-09-07T08:34:03.6054261Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:34:03.6054603Z loading model: 0it [00:14, ?it/s] 2025-09-07T08:34:03.6666779Z cpu eval hf_GPT2_large 2025-09-07T08:34:04.0591428Z pass_due_to_skip 2025-09-07T08:34:04.0595784Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:34:06.4948548Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:34:06.4949402Z import pynvml # type: ignore[import] 2025-09-07T08:34:08.5082966Z 2025-09-07T08:34:12.5491440Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:34:12.5491805Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:34:12.5783022Z cpu eval hf_Longformer 2025-09-07T08:34:14.2844242Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:34:15.0234530Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:34:15.8091895Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:34:16.9129326Z ERROR:common: 2025-09-07T08:34:16.9129937Z Traceback (most recent call last): 2025-09-07T08:34:16.9130316Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 2320, in check_accuracy 2025-09-07T08:34:16.9130695Z optimized_model_iter_fn = optimize_ctx( 2025-09-07T08:34:16.9131022Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 1523, in export 2025-09-07T08:34:16.9131340Z ep = torch.export.export( 2025-09-07T08:34:16.9131708Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 311, in export 2025-09-07T08:34:16.9132069Z raise e 2025-09-07T08:34:16.9132372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 277, in export 2025-09-07T08:34:16.9132693Z return _export( 2025-09-07T08:34:16.9132987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:34:16.9133299Z raise e 2025-09-07T08:34:16.9133663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:34:16.9133995Z ep = fn(*args, **kwargs) 2025-09-07T08:34:16.9134345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:34:16.9134702Z return fn(*args, **kwargs) 2025-09-07T08:34:16.9135108Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2255, in _export 2025-09-07T08:34:16.9135434Z ep = _export_for_training( 2025-09-07T08:34:16.9135739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:34:16.9136053Z raise e 2025-09-07T08:34:16.9136324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:34:16.9136640Z ep = fn(*args, **kwargs) 2025-09-07T08:34:16.9136958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:34:16.9137308Z return fn(*args, **kwargs) 2025-09-07T08:34:16.9137704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2071, in _export_for_training 2025-09-07T08:34:16.9138068Z export_artifact = export_func( 2025-09-07T08:34:16.9138401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1415, in _strict_export 2025-09-07T08:34:16.9138758Z gm_torch_level = _export_to_torch_ir( 2025-09-07T08:34:16.9139116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 812, in _export_to_torch_ir 2025-09-07T08:34:16.9139488Z gm_torch_level, _ = torch._dynamo.export( 2025-09-07T08:34:16.9139837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 2002, in inner 2025-09-07T08:34:16.9140173Z result_traced = opt_f(*args, **kwargs) 2025-09-07T08:34:16.9140515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 414, in __call__ 2025-09-07T08:34:16.9140862Z return super().__call__(*args, **kwargs) 2025-09-07T08:34:16.9141242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl 2025-09-07T08:34:16.9141615Z return self._call_impl(*args, **kwargs) 2025-09-07T08:34:16.9141958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:34:16.9142305Z return forward_call(*args, **kwargs) 2025-09-07T08:34:16.9142659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 841, in compile_wrapper 2025-09-07T08:34:16.9143151Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-09-07T08:34:16.9143715Z torch._dynamo.exc.UserError: Consider annotating your code using torch._check*(). Could not guard on data-dependent expression Eq(u0, 1) (unhinted: Eq(u0, 1)). (Size-like symbols: none) 2025-09-07T08:34:16.9144186Z 2025-09-07T08:34:16.9144745Z consider using data-dependent friendly APIs such as guard_or_false, guard_or_true and statically_known_trueCaused by: if is_global_attn: # transformers/models/longformer/modeling_longformer.py:554 in forward (_dynamo/variables/tensor.py:1435 in evaluate_expr) 2025-09-07T08:34:16.9145475Z For more information, run with TORCH_LOGS="dynamic" 2025-09-07T08:34:16.9145817Z For extended logs when we create symbols, also add TORCHDYNAMO_EXTENDED_DEBUG_CREATE_SYMBOL="u0" 2025-09-07T08:34:16.9146228Z If you suspect the guard was triggered from C++, add TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:34:16.9146759Z For more debugging help, see https://docs.google.com/document/d/1HSuTTVvYH1pTew89Rtpeu84Ht3nQEFTYhAX3Ypa_xJs/edit?usp=sharing 2025-09-07T08:34:16.9147102Z 2025-09-07T08:34:16.9147176Z User Stack (most recent call last): 2025-09-07T08:34:16.9147379Z (snipped, see stack below for prefix) 2025-09-07T08:34:16.9147856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1703, in forward 2025-09-07T08:34:16.9148293Z outputs = self.longformer( 2025-09-07T08:34:16.9148637Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:34:16.9148988Z return forward_call(*args, **kwargs) 2025-09-07T08:34:16.9149458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1600, in forward 2025-09-07T08:34:16.9149878Z encoder_outputs = self.encoder( 2025-09-07T08:34:16.9150224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:34:16.9150582Z return forward_call(*args, **kwargs) 2025-09-07T08:34:16.9150994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in forward 2025-09-07T08:34:16.9151412Z layer_outputs = layer_module( 2025-09-07T08:34:16.9151749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:34:16.9152093Z return forward_call(*args, **kwargs) 2025-09-07T08:34:16.9152504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:34:16.9152916Z self_attn_outputs = self.attention( 2025-09-07T08:34:16.9153256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:34:16.9153610Z return forward_call(*args, **kwargs) 2025-09-07T08:34:16.9154011Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:34:16.9154417Z self_outputs = self.self( 2025-09-07T08:34:16.9154748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:34:16.9155097Z return forward_call(*args, **kwargs) 2025-09-07T08:34:16.9155511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 554, in forward 2025-09-07T08:34:16.9155917Z if is_global_attn: 2025-09-07T08:34:16.9156015Z 2025-09-07T08:34:16.9156126Z For C++ stack trace, run with TORCHDYNAMO_EXTENDED_DEBUG_CPP=1 2025-09-07T08:34:16.9156586Z For more information about this error, see: https://pytorch.org/docs/main/generated/exportdb/index.html#constrain-as-size-example 2025-09-07T08:34:16.9156927Z 2025-09-07T08:34:16.9157028Z from user code: 2025-09-07T08:34:16.9157415Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1703, in forward 2025-09-07T08:34:16.9157831Z outputs = self.longformer( 2025-09-07T08:34:16.9158154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:34:16.9158540Z return forward_call(*args, **kwargs) 2025-09-07T08:34:16.9158953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1600, in forward 2025-09-07T08:34:16.9159373Z encoder_outputs = self.encoder( 2025-09-07T08:34:16.9159718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:34:16.9160071Z return forward_call(*args, **kwargs) 2025-09-07T08:34:16.9160485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in forward 2025-09-07T08:34:16.9160906Z layer_outputs = layer_module( 2025-09-07T08:34:16.9161239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:34:16.9161576Z return forward_call(*args, **kwargs) 2025-09-07T08:34:16.9162020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T08:34:16.9162443Z self_attn_outputs = self.attention( 2025-09-07T08:34:16.9162788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:34:16.9163130Z return forward_call(*args, **kwargs) 2025-09-07T08:34:16.9163559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T08:34:16.9163975Z self_outputs = self.self( 2025-09-07T08:34:16.9164299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:34:16.9164641Z return forward_call(*args, **kwargs) 2025-09-07T08:34:16.9165043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 554, in forward 2025-09-07T08:34:16.9165448Z if is_global_attn: 2025-09-07T08:34:16.9165551Z 2025-09-07T08:34:16.9165901Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-09-07T08:34:16.9166308Z 2025-09-07T08:34:16.9166311Z 2025-09-07T08:34:16.9166812Z The error above occurred when calling torch.export.export. If you would like to view some more information about this error, and get a list of all other errors that may occur in your export call, you can replace your `export()` call with `draft_export()`. 2025-09-07T08:34:16.9167496Z TorchDynamo optimized model failed to run because of following error 2025-09-07T08:34:17.0739065Z fail_to_run 2025-09-07T08:34:17.0739424Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:34:18.7522610Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:34:18.7523499Z import pynvml # type: ignore[import] 2025-09-07T08:34:20.7668288Z 2025-09-07T08:34:22.1438843Z loading model: 0it [00:00, ?it/s]Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T08:34:22.1440589Z WARNING:transformers.models.reformer.modeling_reformer:Passing a tuple of `past_key_values` is deprecated and will be removed in Transformers v4.58.0. You should pass an instance of `ReformerDynamicCache` instead, e.g. `past_key_values=ReformerDynamicCache.from_legacy_cache(past_key_values)`. 2025-09-07T08:34:22.9080436Z 2025-09-07T08:34:22.9080942Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:34:22.9127213Z cpu eval hf_Reformer 2025-09-07T08:34:23.4240126Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:34:24.9017676Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:34:26.3721818Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:34:27.1854901Z class GraphModule(torch.nn.Module): 2025-09-07T08:34:27.1855231Z def forward(self, L_input_ids_: "i64[4, 2048][2048, 1]cpu"): 2025-09-07T08:34:27.1855478Z l_input_ids_ = L_input_ids_ 2025-09-07T08:34:27.1855672Z 2025-09-07T08:34:27.1856320Z # File: /opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py:194 in _get_least_common_mult_chunk_len, code: return np.lcm(config.lsh_attn_chunk_length, config.local_attn_chunk_length) 2025-09-07T08:34:27.1857469Z least_common_mult_chunk_length: "i64[][]cpu" = torch__dynamo_utils_wrapped_lcm(64, 64) 2025-09-07T08:34:27.1857816Z 2025-09-07T08:34:27.1858310Z # File: /opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py:2099 in forward, code: input_shape[-1] % least_common_mult_chunk_length != 0 2025-09-07T08:34:27.1859136Z wrapped_mod: "i64[][]cpu" = torch__dynamo_utils_wrapped_mod(2048, least_common_mult_chunk_length); least_common_mult_chunk_length = None 2025-09-07T08:34:27.1859651Z wrapped_ne: "b8[][]cpu" = torch__dynamo_utils_wrapped_ne(wrapped_mod, 0); wrapped_mod = wrapped_ne = None 2025-09-07T08:34:27.1859968Z 2025-09-07T08:34:27.1860122Z class GraphModule(torch.nn.Module): 2025-09-07T08:34:27.1860364Z def forward(self, L_input_ids_: "i64[4, 2048][2048, 1]cpu"): 2025-09-07T08:34:27.1860592Z l_input_ids_ = L_input_ids_ 2025-09-07T08:34:27.1860767Z 2025-09-07T08:34:27.1861375Z # File: /opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py:194 in _get_least_common_mult_chunk_len, code: return np.lcm(config.lsh_attn_chunk_length, config.local_attn_chunk_length) 2025-09-07T08:34:27.1862089Z least_common_mult_chunk_length: "i64[][]cpu" = torch__dynamo_utils_wrapped_lcm(64, 64) 2025-09-07T08:34:27.1862366Z 2025-09-07T08:34:27.1862838Z # File: /opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py:2099 in forward, code: input_shape[-1] % least_common_mult_chunk_length != 0 2025-09-07T08:34:27.1863548Z wrapped_mod: "i64[][]cpu" = torch__dynamo_utils_wrapped_mod(2048, least_common_mult_chunk_length); least_common_mult_chunk_length = None 2025-09-07T08:34:27.1864055Z wrapped_ne: "b8[][]cpu" = torch__dynamo_utils_wrapped_ne(wrapped_mod, 0); wrapped_mod = wrapped_ne = None 2025-09-07T08:34:27.1864355Z 2025-09-07T08:34:27.1864485Z ERROR:common: 2025-09-07T08:34:27.1864635Z Traceback (most recent call last): 2025-09-07T08:34:27.1864966Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 2320, in check_accuracy 2025-09-07T08:34:27.1865300Z optimized_model_iter_fn = optimize_ctx( 2025-09-07T08:34:27.1865612Z File "/var/lib/jenkins/workspace/benchmarks/dynamo/common.py", line 1523, in export 2025-09-07T08:34:27.1865910Z ep = torch.export.export( 2025-09-07T08:34:27.1866250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 311, in export 2025-09-07T08:34:27.1866569Z raise e 2025-09-07T08:34:27.1866844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/__init__.py", line 277, in export 2025-09-07T08:34:27.1867258Z return _export( 2025-09-07T08:34:27.1867550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:34:27.1867868Z raise e 2025-09-07T08:34:27.1868140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:34:27.1868557Z ep = fn(*args, **kwargs) 2025-09-07T08:34:27.1868897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:34:27.1869245Z return fn(*args, **kwargs) 2025-09-07T08:34:27.1869552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2255, in _export 2025-09-07T08:34:27.1869876Z ep = _export_for_training( 2025-09-07T08:34:27.1870178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1163, in wrapper 2025-09-07T08:34:27.1870484Z raise e 2025-09-07T08:34:27.1870758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1129, in wrapper 2025-09-07T08:34:27.1871075Z ep = fn(*args, **kwargs) 2025-09-07T08:34:27.1871400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/exported_program.py", line 124, in wrapper 2025-09-07T08:34:27.1871755Z return fn(*args, **kwargs) 2025-09-07T08:34:27.1872127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 2071, in _export_for_training 2025-09-07T08:34:27.1872501Z export_artifact = export_func( 2025-09-07T08:34:27.1872856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 1415, in _strict_export 2025-09-07T08:34:27.1873216Z gm_torch_level = _export_to_torch_ir( 2025-09-07T08:34:27.1873605Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/export/_trace.py", line 812, in _export_to_torch_ir 2025-09-07T08:34:27.1873983Z gm_torch_level, _ = torch._dynamo.export( 2025-09-07T08:34:27.1874329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 2002, in inner 2025-09-07T08:34:27.1874680Z result_traced = opt_f(*args, **kwargs) 2025-09-07T08:34:27.1875024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 414, in __call__ 2025-09-07T08:34:27.1875376Z return super().__call__(*args, **kwargs) 2025-09-07T08:34:27.1875757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl 2025-09-07T08:34:27.1876137Z return self._call_impl(*args, **kwargs) 2025-09-07T08:34:27.1876491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:34:27.1876841Z return forward_call(*args, **kwargs) 2025-09-07T08:34:27.1877206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 841, in compile_wrapper 2025-09-07T08:34:27.1877620Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-09-07T08:34:27.1877923Z torch._dynamo.exc.Unsupported: Data-dependent branching 2025-09-07T08:34:27.1878357Z Explanation: Detected data-dependent branching (e.g. `if my_tensor.sum() > 0:`). Dynamo does not support tracing dynamic control flow. 2025-09-07T08:34:27.1878945Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-09-07T08:34:27.1879390Z Hint: Use `torch.cond` to express dynamic control flow. 2025-09-07T08:34:27.1879556Z 2025-09-07T08:34:27.1879695Z Developer debug context: attempted to jump with NumpyNdarrayVariable() 2025-09-07T08:34:27.1879894Z 2025-09-07T08:34:27.1880184Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0170.html 2025-09-07T08:34:27.1880503Z 2025-09-07T08:34:27.1880568Z from user code: 2025-09-07T08:34:27.1880992Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2481, in forward 2025-09-07T08:34:27.1881416Z reformer_outputs = self.reformer( 2025-09-07T08:34:27.1881766Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl 2025-09-07T08:34:27.1882167Z return forward_call(*args, **kwargs) 2025-09-07T08:34:27.1882573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2099, in forward 2025-09-07T08:34:27.1883024Z input_shape[-1] % least_common_mult_chunk_length != 0 2025-09-07T08:34:27.1883188Z 2025-09-07T08:34:27.1883536Z Set TORCHDYNAMO_VERBOSE=1 for the internal stack trace (please do this especially if you're reporting a bug to PyTorch). For even more developer context, set TORCH_LOGS="+dynamo" 2025-09-07T08:34:27.1883944Z 2025-09-07T08:34:27.1884080Z TorchDynamo optimized model failed to run because of following error 2025-09-07T08:34:27.2618279Z fail_to_run 2025-09-07T08:34:27.2618622Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:34:28.5499135Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:34:28.5500044Z import pynvml # type: ignore[import] 2025-09-07T08:34:30.5738501Z 2025-09-07T08:34:36.3887173Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:34:36.3887652Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:34:36.4218471Z cpu eval hf_Roberta_base 2025-09-07T08:34:38.3125015Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:34:38.8458510Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:34:39.3812901Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:34:47.7760955Z pass 2025-09-07T08:34:47.7761381Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:34:50.1234167Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:34:50.1235043Z import pynvml # type: ignore[import] 2025-09-07T08:34:52.1410528Z 2025-09-07T08:34:55.0227380Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:34:55.0227729Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:34:55.0447517Z cpu eval hf_T5 2025-09-07T08:34:56.0469830Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:34:56.5111562Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:34:56.9612658Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:35:04.7318957Z pass 2025-09-07T08:35:04.7319396Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:35:06.4315574Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:35:06.4316475Z import pynvml # type: ignore[import] 2025-09-07T08:35:08.4407438Z 2025-09-07T08:35:17.8935671Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:35:17.8936007Z loading model: 0it [00:09, ?it/s] 2025-09-07T08:35:17.9339282Z cpu eval hf_T5_base 2025-09-07T08:35:32.3534128Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:35:38.8444558Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:35:45.2895648Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:36:11.6535767Z pass 2025-09-07T08:36:11.6536167Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:36:16.9724855Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:36:16.9725959Z import pynvml # type: ignore[import] 2025-09-07T08:36:18.9850592Z 2025-09-07T08:36:31.5177764Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:36:31.5178325Z loading model: 0it [00:12, ?it/s] 2025-09-07T08:36:31.6025014Z cpu eval hf_T5_large 2025-09-07T08:36:32.0029131Z pass_due_to_skip 2025-09-07T08:36:32.0032770Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:36:34.3331375Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:36:34.3332258Z import pynvml # type: ignore[import] 2025-09-07T08:36:36.3476945Z 2025-09-07T08:36:42.5975094Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:36:42.5975399Z loading model: 0it [00:06, ?it/s] 2025-09-07T08:36:42.6358718Z cpu eval hf_distil_whisper 2025-09-07T08:36:50.0496551Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:36:52.1460243Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:36:54.2316303Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:37:04.7195052Z pass 2025-09-07T08:37:04.7195467Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:37:06.6759844Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:37:06.6760717Z import pynvml # type: ignore[import] 2025-09-07T08:37:08.6983645Z 2025-09-07T08:37:08.8523134Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:37:08.8523431Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:37:08.8523655Z cpu eval lennard_jones 2025-09-07T08:37:08.8534986Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:37:08.8560465Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:37:08.8575307Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:37:13.5120166Z pass 2025-09-07T08:37:13.5120574Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:37:14.2201053Z accuracy pass_rate=65.22% 2025-09-07T08:37:14.2205102Z calls_captured gmean=0.00x mean=263.391x 2025-09-07T08:37:14.2207477Z unique_graphs gmean=0.00x mean=0.652x 2025-09-07T08:37:14.2209564Z graph_breaks gmean=0.00x mean=0.000x 2025-09-07T08:37:14.2211690Z unique_graph_breaks gmean=0.00x mean=0.000x 2025-09-07T08:37:14.2213877Z autograd_captures gmean=0.00x mean=0.000x 2025-09-07T08:37:14.2216052Z autograd_compiles gmean=0.00x mean=0.000x 2025-09-07T08:37:14.2217962Z cudagraph_skips gmean=0.00x mean=0.000x 2025-09-07T08:37:14.2218662Z compilation_latency mean=4.096 seconds 2025-09-07T08:37:14.7479623Z + taskset -c 0-94 python benchmarks/dynamo/torchbench.py --accuracy --no-translation-validation --inference --bfloat16 --export-aot-inductor --disable-cudagraphs --device cpu --total-partitions 4 --partition-id 1 --output /var/lib/jenkins/workspace/test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.csv 2025-09-07T08:37:15.1433620Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:37:15.1434475Z import pynvml # type: ignore[import] 2025-09-07T08:37:17.5924593Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:37:17.5926127Z import pynvml # type: ignore[import] 2025-09-07T08:37:19.6130870Z 2025-09-07T08:37:24.8780882Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:37:24.8781237Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:37:24.9059463Z cpu eval dlrm 2025-09-07T08:37:25.0430125Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:37:25.0712867Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:37:25.0970766Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:37:44.6408023Z pass 2025-09-07T08:37:44.6410583Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:37:46.5723338Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:37:46.5724230Z import pynvml # type: ignore[import] 2025-09-07T08:37:48.5861028Z 2025-09-07T08:37:49.0211412Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:37:49.0211692Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:37:49.0243278Z cpu eval functorch_dp_cifar10 2025-09-07T08:37:49.0913700Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:37:49.1185198Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:37:49.1412373Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:37:59.6697842Z pass 2025-09-07T08:37:59.6700190Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:38:01.4273749Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:38:01.4274647Z import pynvml # type: ignore[import] 2025-09-07T08:38:03.4458559Z 2025-09-07T08:38:03.6402649Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:38:03.6403314Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:38:03.6408132Z cpu eval functorch_maml_omniglot 2025-09-07T08:38:03.6503338Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:38:03.6581027Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:38:03.6629217Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:38:11.5172422Z pass 2025-09-07T08:38:11.5174714Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:38:13.0291111Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:38:13.0292006Z import pynvml # type: ignore[import] 2025-09-07T08:38:15.0278916Z 2025-09-07T08:38:16.6958340Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:38:16.6958835Z loading model: 0it [00:01, ?it/s] 2025-09-07T08:38:16.7006164Z cpu eval hf_Albert 2025-09-07T08:38:17.6140975Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:38:17.8557987Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:38:18.0830639Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:38:30.0639391Z pass 2025-09-07T08:38:30.0639800Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:38:32.1960880Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:38:32.1961766Z import pynvml # type: ignore[import] 2025-09-07T08:38:34.2104656Z 2025-09-07T08:38:38.4429828Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:38:38.4430794Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:38:38.4690774Z cpu eval hf_Bart 2025-09-07T08:38:39.8095924Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:38:40.1578313Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:38:40.5060568Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:38:56.9680090Z pass 2025-09-07T08:38:56.9680476Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:38:59.5200055Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:38:59.5200933Z import pynvml # type: ignore[import] 2025-09-07T08:39:01.5420460Z 2025-09-07T08:39:04.6738714Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:39:04.6739040Z loading model: 0it [00:03, ?it/s] 2025-09-07T08:39:04.6973428Z cpu eval hf_Bert 2025-09-07T08:39:05.7895980Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:39:06.0766254Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:39:06.3523517Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:39:20.0822608Z pass 2025-09-07T08:39:20.0823040Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:39:22.3920171Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:39:22.3921406Z import pynvml # type: ignore[import] 2025-09-07T08:39:24.4025691Z 2025-09-07T08:39:29.8114124Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:39:29.8114848Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:39:29.8500087Z cpu eval hf_Bert_large 2025-09-07T08:39:32.1164637Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:39:32.7673704Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:39:33.4072239Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:39:53.9832572Z pass 2025-09-07T08:39:53.9832977Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:39:56.7302105Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:39:56.7303021Z import pynvml # type: ignore[import] 2025-09-07T08:39:58.7465171Z 2025-09-07T08:40:00.1488689Z loading model: 0it [00:00, ?it/s]BigBirdForMaskedLM has generative capabilities, as `prepare_inputs_for_generation` is explicitly defined. However, it doesn't directly inherit from `GenerationMixin`. From 👉v4.50👈 onwards, `PreTrainedModel` will NOT inherit from `GenerationMixin`, and this model will lose the ability to call `generate` and other related functions. 2025-09-07T08:40:00.1490355Z - If you're using `trust_remote_code=True`, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes 2025-09-07T08:40:00.1491195Z - If you are the owner of the model architecture code, please modify your model class such that it inherits from `GenerationMixin` (after `PreTrainedModel`, otherwise you'll get an exception). 2025-09-07T08:40:00.1491848Z - If you are not the owner of the model architecture class, please contact the model code owner to update it. 2025-09-07T08:40:00.1493151Z WARNING:transformers.modeling_utils:BigBirdForMaskedLM has generative capabilities, as `prepare_inputs_for_generation` is explicitly defined. However, it doesn't directly inherit from `GenerationMixin`. From 👉v4.50👈 onwards, `PreTrainedModel` will NOT inherit from `GenerationMixin`, and this model will lose the ability to call `generate` and other related functions. 2025-09-07T08:40:00.1494361Z - If you're using `trust_remote_code=True`, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes 2025-09-07T08:40:00.1495173Z - If you are the owner of the model architecture code, please modify your model class such that it inherits from `GenerationMixin` (after `PreTrainedModel`, otherwise you'll get an exception). 2025-09-07T08:40:00.1495805Z - If you are not the owner of the model architecture class, please contact the model code owner to update it. 2025-09-07T08:40:01.9783179Z Input ids are automatically padded from 819 to 832 to be a multiple of `config.block_size`: 64 2025-09-07T08:40:01.9783965Z WARNING:transformers.models.big_bird.modeling_big_bird:Input ids are automatically padded from 819 to 832 to be a multiple of `config.block_size`: 64 2025-09-07T08:40:03.8081202Z 2025-09-07T08:40:03.8082237Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:40:03.8373657Z cpu eval hf_BigBird 2025-09-07T08:40:06.3139571Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:40:09.3269645Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:40:12.4237063Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:41:17.3992174Z pass 2025-09-07T08:41:17.3992583Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:41:21.4516938Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:41:21.4518040Z import pynvml # type: ignore[import] 2025-09-07T08:41:23.4740724Z 2025-09-07T08:41:26.0897512Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:41:26.0897807Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:41:26.1007696Z cpu eval hf_DistilBert 2025-09-07T08:41:26.6822473Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:41:26.8458267Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:41:27.0005720Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:41:38.1747812Z pass 2025-09-07T08:41:38.1748556Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:41:40.3640473Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:41:40.3641712Z import pynvml # type: ignore[import] 2025-09-07T08:41:42.3785153Z 2025-09-07T08:41:47.0203374Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:41:47.0203707Z loading model: 0it [00:04, ?it/s] 2025-09-07T08:41:47.0382531Z cpu eval hf_GPT2 2025-09-07T08:41:47.5880209Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:41:47.7514752Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:41:47.9073039Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:42:02.2466757Z pass 2025-09-07T08:42:02.2467162Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:42:04.6397283Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:42:04.6398177Z import pynvml # type: ignore[import] 2025-09-07T08:42:06.6495481Z 2025-09-07T08:42:20.8970973Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:42:20.8971474Z loading model: 0it [00:14, ?it/s] 2025-09-07T08:42:20.9576290Z cpu eval hf_GPT2_large 2025-09-07T08:42:21.3455799Z pass_due_to_skip 2025-09-07T08:42:21.3459990Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:42:23.8005142Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:42:23.8006036Z import pynvml # type: ignore[import] 2025-09-07T08:42:25.8119337Z 2025-09-07T08:42:31.5946531Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:42:31.5947266Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:42:31.6231513Z cpu eval hf_Roberta_base 2025-09-07T08:42:33.5253074Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:42:34.0613892Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:42:34.5985464Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:42:53.9655108Z pass 2025-09-07T08:42:53.9655495Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:42:57.2914229Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:42:57.2915119Z import pynvml # type: ignore[import] 2025-09-07T08:42:59.2980620Z 2025-09-07T08:43:02.1408739Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:43:02.1409037Z loading model: 0it [00:02, ?it/s] 2025-09-07T08:43:02.1620730Z cpu eval hf_T5 2025-09-07T08:43:03.1440566Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:43:03.6116623Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:43:04.0672554Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:43:21.0515034Z pass 2025-09-07T08:43:21.0519878Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:43:23.5164239Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:43:23.5165250Z import pynvml # type: ignore[import] 2025-09-07T08:43:25.5263168Z 2025-09-07T08:43:35.1524535Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:43:35.1524856Z loading model: 0it [00:09, ?it/s] 2025-09-07T08:43:35.1940687Z cpu eval hf_T5_base 2025-09-07T08:43:49.5763628Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:43:56.0758067Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:44:02.5616491Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:44:50.4827110Z pass 2025-09-07T08:44:50.4827540Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:44:56.9025179Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:44:56.9026112Z import pynvml # type: ignore[import] 2025-09-07T08:44:58.9137030Z 2025-09-07T08:45:11.2923345Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:45:11.2923669Z loading model: 0it [00:12, ?it/s] 2025-09-07T08:45:11.3761923Z cpu eval hf_T5_large 2025-09-07T08:45:11.7720489Z pass_due_to_skip 2025-09-07T08:45:11.7723274Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:45:14.1173521Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:45:14.1174391Z import pynvml # type: ignore[import] 2025-09-07T08:45:16.1299197Z 2025-09-07T08:45:22.3460909Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:45:22.3461184Z loading model: 0it [00:06, ?it/s] 2025-09-07T08:45:22.3845095Z cpu eval hf_distil_whisper 2025-09-07T08:45:29.7897173Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:45:31.8852909Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:45:34.0392569Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:45:58.2335811Z pass 2025-09-07T08:45:58.2336214Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:46:01.1367762Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:46:01.1368733Z import pynvml # type: ignore[import] 2025-09-07T08:46:03.1393524Z 2025-09-07T08:46:03.2922704Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:46:03.2922976Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:46:03.2924365Z cpu eval lennard_jones 2025-09-07T08:46:03.2937498Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:46:03.2963583Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:46:03.2978501Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:46:10.2483856Z pass 2025-09-07T08:46:10.2486530Z WARNING:common:Trying to call the empty_gpu_cache for device: cpu, which is not in list [cuda, xpu] 2025-09-07T08:46:11.3323273Z accuracy pass_rate=88.24% 2025-09-07T08:46:11.3326216Z calls_captured gmean=0.00x mean=0.000x 2025-09-07T08:46:11.3328596Z unique_graphs gmean=0.00x mean=0.000x 2025-09-07T08:46:11.3330727Z graph_breaks gmean=0.00x mean=0.000x 2025-09-07T08:46:11.3332837Z unique_graph_breaks gmean=0.00x mean=0.000x 2025-09-07T08:46:11.3334961Z autograd_captures gmean=0.00x mean=0.000x 2025-09-07T08:46:11.3337172Z autograd_compiles gmean=0.00x mean=0.000x 2025-09-07T08:46:11.3339384Z cudagraph_skips gmean=0.00x mean=0.000x 2025-09-07T08:46:11.3340416Z compilation_latency mean=0.000 seconds 2025-09-07T08:46:11.8628645Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *maxautotune-true* ]] 2025-09-07T08:46:11.8629454Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *cudagraphs_low_precision-true* ]] 2025-09-07T08:46:11.8629962Z + for target in "${targets[@]}" 2025-09-07T08:46:11.8630159Z + target_flag=('--performance') 2025-09-07T08:46:11.8630339Z + local target_flag 2025-09-07T08:46:11.8630521Z + [[ performance == \p\e\r\f\o\r\m\a\n\c\e ]] 2025-09-07T08:46:11.8630734Z + target_flag+=(--cold-start-latency) 2025-09-07T08:46:11.8631188Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *freezing-true* ]] 2025-09-07T08:46:11.8631860Z + [[ training-false-inference-true-default-true-dynamic-true-cppwrapper-true-aotinductor-true == *default-true* ]] 2025-09-07T08:46:11.8633040Z + taskset -c 0-94 python benchmarks/dynamo/torchbench.py --performance --cold-start-latency --inference --bfloat16 --backend inductor --disable-cudagraphs --device cpu --total-partitions 4 --partition-id 1 --output /var/lib/jenkins/workspace/test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_x86_zen_performance.csv 2025-09-07T08:46:12.2614513Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:46:12.2615897Z import pynvml # type: ignore[import] 2025-09-07T08:46:14.7009183Z /opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you. 2025-09-07T08:46:14.7010386Z import pynvml # type: ignore[import] 2025-09-07T08:46:16.7219066Z 2025-09-07T08:46:22.3804991Z loading model: 0it [00:00, ?it/s] 2025-09-07T08:46:22.3805598Z loading model: 0it [00:05, ?it/s] 2025-09-07T08:46:22.4012045Z cpu eval dlrm 2025-09-07T08:46:35.3560380Z 2025-09-07T08:46:35.4917561Z running benchmark: 0% 0/30 [00:00 2025-09-07T09:03:10.6293993Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T09:03:10.6294189Z 2025-09-07T09:03:10.6294290Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6294753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6295191Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6295639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T09:03:10.6296138Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T09:03:10.6296634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T09:03:10.6297133Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T09:03:10.6297354Z 2025-09-07T09:03:10.6297458Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6298007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6298442Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6298859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T09:03:10.6299382Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T09:03:10.6299903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T09:03:10.6300396Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T09:03:10.6300588Z 2025-09-07T09:03:10.6300684Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6301147Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6301577Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6301990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T09:03:10.6302469Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T09:03:10.6302968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T09:03:10.6303692Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T09:03:10.6303882Z 2025-09-07T09:03:10.6303982Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6304435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6304860Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6305267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T09:03:10.6305902Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6306564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T09:03:10.6307124Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6307300Z 2025-09-07T09:03:10.6307402Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6307873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6308309Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6308736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T09:03:10.6309214Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6309705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T09:03:10.6310184Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6310366Z 2025-09-07T09:03:10.6310508Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6310968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6311407Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6311857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T09:03:10.6312343Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6312822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T09:03:10.6313300Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6313467Z 2025-09-07T09:03:10.6313563Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6314020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6314441Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6314850Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T09:03:10.6315318Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6315792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T09:03:10.6316271Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6316436Z 2025-09-07T09:03:10.6316524Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6316973Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T09:03:10.6317417Z logits = self.prob_head(feat_concat) 2025-09-07T09:03:10.6317541Z 2025-09-07T09:03:10.6317635Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6318083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T09:03:10.6318507Z logits = self.prob_head(feat_concat) 2025-09-07T09:03:10.6318636Z 2025-09-07T09:03:10.6318725Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6319174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T09:03:10.6319655Z logits = self.prob_head(feat_concat) 2025-09-07T09:03:10.6319779Z 2025-09-07T09:03:10.6319878Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6320341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T09:03:10.6320813Z logits = self.prob_head(feat_concat) 2025-09-07T09:03:10.6320938Z 2025-09-07T09:03:10.6321026Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6321478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6321911Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6322246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6322575Z x = module(x) 2025-09-07T09:03:10.6322668Z 2025-09-07T09:03:10.6322758Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6323208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6323635Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6323985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6324309Z x = module(x) 2025-09-07T09:03:10.6324390Z 2025-09-07T09:03:10.6324486Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6324988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6325412Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6325736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6326059Z x = module(x) 2025-09-07T09:03:10.6326137Z 2025-09-07T09:03:10.6326234Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6326678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6327106Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6327420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6327741Z x = module(x) 2025-09-07T09:03:10.6328037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T09:03:10.6328364Z out = self.conv1(x) 2025-09-07T09:03:10.6328469Z 2025-09-07T09:03:10.6328558Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6329009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6329437Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6329774Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6330102Z x = module(x) 2025-09-07T09:03:10.6330423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T09:03:10.6330768Z out = self.relu(out) 2025-09-07T09:03:10.6330871Z 2025-09-07T09:03:10.6330976Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6331438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6331876Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6332208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6332575Z x = module(x) 2025-09-07T09:03:10.6332877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T09:03:10.6333209Z out = self.conv2(out) 2025-09-07T09:03:10.6333319Z 2025-09-07T09:03:10.6333412Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6333906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6334343Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6334664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6334983Z x = module(x) 2025-09-07T09:03:10.6335279Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T09:03:10.6335609Z out = self.relu(out) 2025-09-07T09:03:10.6335704Z 2025-09-07T09:03:10.6335803Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6336256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6336689Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6337045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6337376Z x = module(x) 2025-09-07T09:03:10.6337712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T09:03:10.6338045Z out = self.conv3(out) 2025-09-07T09:03:10.6338150Z 2025-09-07T09:03:10.6338243Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6338732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6339166Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6339486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6339802Z x = module(x) 2025-09-07T09:03:10.6340089Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 158, in forward 2025-09-07T09:03:10.6340435Z identity = self.downsample(x) 2025-09-07T09:03:10.6340555Z 2025-09-07T09:03:10.6340646Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6341088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6341509Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6341819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6342129Z x = module(x) 2025-09-07T09:03:10.6342404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T09:03:10.6342724Z out = self.conv1(x) 2025-09-07T09:03:10.6342818Z 2025-09-07T09:03:10.6342905Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6343344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6343768Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6344071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6344382Z x = module(x) 2025-09-07T09:03:10.6344657Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T09:03:10.6344979Z out = self.relu(out) 2025-09-07T09:03:10.6345074Z 2025-09-07T09:03:10.6345162Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6345597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6346078Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6346388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6346697Z x = module(x) 2025-09-07T09:03:10.6347019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T09:03:10.6347346Z out = self.conv2(out) 2025-09-07T09:03:10.6347445Z 2025-09-07T09:03:10.6347531Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6347969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6348384Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6348702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6349022Z x = module(x) 2025-09-07T09:03:10.6349312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T09:03:10.6349635Z out = self.relu(out) 2025-09-07T09:03:10.6349731Z 2025-09-07T09:03:10.6349824Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6350309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6350737Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6351051Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6351364Z x = module(x) 2025-09-07T09:03:10.6351676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T09:03:10.6352003Z out = self.conv3(out) 2025-09-07T09:03:10.6352102Z 2025-09-07T09:03:10.6352194Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6352638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6353055Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6353362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6353676Z x = module(x) 2025-09-07T09:03:10.6353953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T09:03:10.6354272Z out = self.relu(out) 2025-09-07T09:03:10.6354365Z 2025-09-07T09:03:10.6354457Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6354908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6355327Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6355642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6355958Z x = module(x) 2025-09-07T09:03:10.6356239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T09:03:10.6356563Z out = self.conv1(x) 2025-09-07T09:03:10.6356659Z 2025-09-07T09:03:10.6356751Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6357192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6357614Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6357925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6358235Z x = module(x) 2025-09-07T09:03:10.6358515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T09:03:10.6358889Z out = self.relu(out) 2025-09-07T09:03:10.6358984Z 2025-09-07T09:03:10.6359074Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6359516Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6359984Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6360294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6360606Z x = module(x) 2025-09-07T09:03:10.6360965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T09:03:10.6361287Z out = self.conv2(out) 2025-09-07T09:03:10.6361388Z 2025-09-07T09:03:10.6361477Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6361915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6362338Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6362642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6362954Z x = module(x) 2025-09-07T09:03:10.6363284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T09:03:10.6363610Z out = self.relu(out) 2025-09-07T09:03:10.6363703Z 2025-09-07T09:03:10.6363789Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6364231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6364690Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6364996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6365315Z x = module(x) 2025-09-07T09:03:10.6365589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T09:03:10.6365909Z out = self.conv3(out) 2025-09-07T09:03:10.6366004Z 2025-09-07T09:03:10.6366090Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6366534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6366949Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6367252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6367562Z x = module(x) 2025-09-07T09:03:10.6367839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T09:03:10.6368162Z out = self.relu(out) 2025-09-07T09:03:10.6368255Z 2025-09-07T09:03:10.6368345Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6368790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6369211Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6369530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6369845Z x = module(x) 2025-09-07T09:03:10.6370122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T09:03:10.6370442Z out = self.conv1(x) 2025-09-07T09:03:10.6370533Z 2025-09-07T09:03:10.6370624Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6371067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6371482Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6371850Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6372161Z x = module(x) 2025-09-07T09:03:10.6372440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T09:03:10.6372763Z out = self.relu(out) 2025-09-07T09:03:10.6372903Z 2025-09-07T09:03:10.6373003Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6373464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6373893Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6374220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6374539Z x = module(x) 2025-09-07T09:03:10.6374830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T09:03:10.6375159Z out = self.conv2(out) 2025-09-07T09:03:10.6375258Z 2025-09-07T09:03:10.6375351Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6375799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6376278Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6376595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6376917Z x = module(x) 2025-09-07T09:03:10.6377209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T09:03:10.6377538Z out = self.relu(out) 2025-09-07T09:03:10.6377721Z 2025-09-07T09:03:10.6377816Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6378268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6378696Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6379016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6379327Z x = module(x) 2025-09-07T09:03:10.6379617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T09:03:10.6379940Z out = self.conv3(out) 2025-09-07T09:03:10.6380039Z 2025-09-07T09:03:10.6380131Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6380574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6380997Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6381309Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6381621Z x = module(x) 2025-09-07T09:03:10.6381901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 158, in forward 2025-09-07T09:03:10.6382234Z identity = self.downsample(x) 2025-09-07T09:03:10.6382350Z 2025-09-07T09:03:10.6382440Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6382887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6383307Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6383617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6383927Z x = module(x) 2025-09-07T09:03:10.6384211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T09:03:10.6384533Z out = self.conv1(x) 2025-09-07T09:03:10.6384627Z 2025-09-07T09:03:10.6384757Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6385200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6385615Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6385926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6386280Z x = module(x) 2025-09-07T09:03:10.6386560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T09:03:10.6386880Z out = self.relu(out) 2025-09-07T09:03:10.6386980Z 2025-09-07T09:03:10.6387068Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6387513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6387931Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6388242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6388553Z x = module(x) 2025-09-07T09:03:10.6388835Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T09:03:10.6389164Z out = self.conv2(out) 2025-09-07T09:03:10.6389298Z 2025-09-07T09:03:10.6389389Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6389830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6390243Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6390617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6390938Z x = module(x) 2025-09-07T09:03:10.6391227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T09:03:10.6391550Z out = self.relu(out) 2025-09-07T09:03:10.6391647Z 2025-09-07T09:03:10.6391736Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6392176Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6392598Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6392907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6393215Z x = module(x) 2025-09-07T09:03:10.6393492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T09:03:10.6393812Z out = self.conv3(out) 2025-09-07T09:03:10.6393906Z 2025-09-07T09:03:10.6393994Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6394433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6394849Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6395156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6395467Z x = module(x) 2025-09-07T09:03:10.6395747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T09:03:10.6396068Z out = self.relu(out) 2025-09-07T09:03:10.6396160Z 2025-09-07T09:03:10.6396246Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6396684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6397099Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6397405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6397761Z x = module(x) 2025-09-07T09:03:10.6398038Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T09:03:10.6398357Z out = self.conv1(x) 2025-09-07T09:03:10.6398447Z 2025-09-07T09:03:10.6398535Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6399012Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6399438Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6399748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6400063Z x = module(x) 2025-09-07T09:03:10.6400343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T09:03:10.6400663Z out = self.relu(out) 2025-09-07T09:03:10.6400766Z 2025-09-07T09:03:10.6400857Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6401306Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6401732Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6402129Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6402462Z x = module(x) 2025-09-07T09:03:10.6402756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T09:03:10.6403256Z out = self.conv2(out) 2025-09-07T09:03:10.6403354Z 2025-09-07T09:03:10.6403450Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6403971Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6404406Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6404727Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6405051Z x = module(x) 2025-09-07T09:03:10.6405338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T09:03:10.6405663Z out = self.relu(out) 2025-09-07T09:03:10.6405761Z 2025-09-07T09:03:10.6405852Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6406307Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6406735Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6407047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6407369Z x = module(x) 2025-09-07T09:03:10.6407656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T09:03:10.6407988Z out = self.conv3(out) 2025-09-07T09:03:10.6408083Z 2025-09-07T09:03:10.6408181Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6408623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6409048Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6409355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6409670Z x = module(x) 2025-09-07T09:03:10.6409951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T09:03:10.6410269Z out = self.relu(out) 2025-09-07T09:03:10.6410365Z 2025-09-07T09:03:10.6410452Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6410897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6411394Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6411717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6412036Z x = module(x) 2025-09-07T09:03:10.6412383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T09:03:10.6412713Z out = self.conv1(x) 2025-09-07T09:03:10.6412810Z 2025-09-07T09:03:10.6412905Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6413353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6413781Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6414099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6414414Z x = module(x) 2025-09-07T09:03:10.6414694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T09:03:10.6415018Z out = self.relu(out) 2025-09-07T09:03:10.6415117Z 2025-09-07T09:03:10.6415207Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6415705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6416134Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6416441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6416765Z x = module(x) 2025-09-07T09:03:10.6417080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T09:03:10.6417413Z out = self.conv2(out) 2025-09-07T09:03:10.6417513Z 2025-09-07T09:03:10.6417654Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6418096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6418520Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6418833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6419151Z x = module(x) 2025-09-07T09:03:10.6419427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T09:03:10.6419751Z out = self.relu(out) 2025-09-07T09:03:10.6419847Z 2025-09-07T09:03:10.6419932Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6420376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6420806Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6421111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6421429Z x = module(x) 2025-09-07T09:03:10.6421707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T09:03:10.6422037Z out = self.conv3(out) 2025-09-07T09:03:10.6422129Z 2025-09-07T09:03:10.6422219Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6422662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6423092Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6423401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6423718Z x = module(x) 2025-09-07T09:03:10.6423993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T09:03:10.6424367Z out = self.relu(out) 2025-09-07T09:03:10.6424461Z 2025-09-07T09:03:10.6424548Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6424996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6425469Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6425776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6426093Z x = module(x) 2025-09-07T09:03:10.6426371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T09:03:10.6426695Z out = self.conv1(x) 2025-09-07T09:03:10.6426789Z 2025-09-07T09:03:10.6426876Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6427321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6427741Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6428054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6428368Z x = module(x) 2025-09-07T09:03:10.6428679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T09:03:10.6429011Z out = self.relu(out) 2025-09-07T09:03:10.6429104Z 2025-09-07T09:03:10.6429193Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6429631Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6430084Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6430389Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6430702Z x = module(x) 2025-09-07T09:03:10.6430981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T09:03:10.6431304Z out = self.conv2(out) 2025-09-07T09:03:10.6431395Z 2025-09-07T09:03:10.6431481Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6431923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6432339Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6432652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6432976Z x = module(x) 2025-09-07T09:03:10.6433255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T09:03:10.6433581Z out = self.relu(out) 2025-09-07T09:03:10.6433671Z 2025-09-07T09:03:10.6433759Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6434208Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6434623Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6434927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6435256Z x = module(x) 2025-09-07T09:03:10.6435534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T09:03:10.6435855Z out = self.conv3(out) 2025-09-07T09:03:10.6435945Z 2025-09-07T09:03:10.6436032Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6436471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6436901Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6437248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6437567Z x = module(x) 2025-09-07T09:03:10.6437841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 158, in forward 2025-09-07T09:03:10.6438228Z identity = self.downsample(x) 2025-09-07T09:03:10.6438352Z 2025-09-07T09:03:10.6438439Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6438877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6439293Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6439604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6439923Z x = module(x) 2025-09-07T09:03:10.6440198Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T09:03:10.6440525Z out = self.conv1(x) 2025-09-07T09:03:10.6440616Z 2025-09-07T09:03:10.6440703Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6441137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6441616Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6441923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6442246Z x = module(x) 2025-09-07T09:03:10.6442520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T09:03:10.6442845Z out = self.relu(out) 2025-09-07T09:03:10.6442975Z 2025-09-07T09:03:10.6443064Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6443502Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6443918Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6444231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6444544Z x = module(x) 2025-09-07T09:03:10.6444830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T09:03:10.6445149Z out = self.conv2(out) 2025-09-07T09:03:10.6445239Z 2025-09-07T09:03:10.6445324Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6445765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6446187Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6446494Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6446804Z x = module(x) 2025-09-07T09:03:10.6447082Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T09:03:10.6447401Z out = self.relu(out) 2025-09-07T09:03:10.6447492Z 2025-09-07T09:03:10.6447585Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6448031Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6448446Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6448755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6449071Z x = module(x) 2025-09-07T09:03:10.6449353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T09:03:10.6449675Z out = self.conv3(out) 2025-09-07T09:03:10.6449767Z 2025-09-07T09:03:10.6449914Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6450363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6450783Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6451091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6451446Z x = module(x) 2025-09-07T09:03:10.6451724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T09:03:10.6452049Z out = self.relu(out) 2025-09-07T09:03:10.6452138Z 2025-09-07T09:03:10.6452228Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6452671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6453086Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6453396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6453709Z x = module(x) 2025-09-07T09:03:10.6453986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T09:03:10.6454310Z out = self.conv1(x) 2025-09-07T09:03:10.6454434Z 2025-09-07T09:03:10.6454521Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6454962Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6455380Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6455724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6456042Z x = module(x) 2025-09-07T09:03:10.6456325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T09:03:10.6456649Z out = self.relu(out) 2025-09-07T09:03:10.6456739Z 2025-09-07T09:03:10.6456828Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6457272Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6457726Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6458037Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6458353Z x = module(x) 2025-09-07T09:03:10.6458634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T09:03:10.6458953Z out = self.conv2(out) 2025-09-07T09:03:10.6459053Z 2025-09-07T09:03:10.6459139Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6459585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6460012Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6460323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6460633Z x = module(x) 2025-09-07T09:03:10.6460916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T09:03:10.6461239Z out = self.relu(out) 2025-09-07T09:03:10.6461328Z 2025-09-07T09:03:10.6461419Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6461876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6462300Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6462615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6463007Z x = module(x) 2025-09-07T09:03:10.6463295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T09:03:10.6463618Z out = self.conv3(out) 2025-09-07T09:03:10.6463718Z 2025-09-07T09:03:10.6463806Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6464305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6464734Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6465047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6465364Z x = module(x) 2025-09-07T09:03:10.6465653Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T09:03:10.6465983Z out = self.relu(out) 2025-09-07T09:03:10.6466073Z 2025-09-07T09:03:10.6466173Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6466617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6467041Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6467397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6467727Z x = module(x) 2025-09-07T09:03:10.6468013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T09:03:10.6468335Z out = self.conv1(x) 2025-09-07T09:03:10.6468433Z 2025-09-07T09:03:10.6468521Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6469009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6469441Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6469760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6470074Z x = module(x) 2025-09-07T09:03:10.6470363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T09:03:10.6470696Z out = self.relu(out) 2025-09-07T09:03:10.6470791Z 2025-09-07T09:03:10.6470888Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6471333Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6471771Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6472091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6472413Z x = module(x) 2025-09-07T09:03:10.6472704Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T09:03:10.6473029Z out = self.conv2(out) 2025-09-07T09:03:10.6473135Z 2025-09-07T09:03:10.6473223Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6473673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6474114Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6474430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6474750Z x = module(x) 2025-09-07T09:03:10.6475036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T09:03:10.6475364Z out = self.relu(out) 2025-09-07T09:03:10.6475455Z 2025-09-07T09:03:10.6475550Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6475994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6476505Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6476833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6477164Z x = module(x) 2025-09-07T09:03:10.6477500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T09:03:10.6477834Z out = self.conv3(out) 2025-09-07T09:03:10.6477940Z 2025-09-07T09:03:10.6478033Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6478495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6478934Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6479251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6479573Z x = module(x) 2025-09-07T09:03:10.6479865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T09:03:10.6480196Z out = self.relu(out) 2025-09-07T09:03:10.6480293Z 2025-09-07T09:03:10.6480390Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6480875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6481306Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6481630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6481958Z x = module(x) 2025-09-07T09:03:10.6482276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T09:03:10.6482611Z out = self.conv1(x) 2025-09-07T09:03:10.6482712Z 2025-09-07T09:03:10.6482807Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6483263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6483692Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6484005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6484326Z x = module(x) 2025-09-07T09:03:10.6484617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T09:03:10.6484950Z out = self.relu(out) 2025-09-07T09:03:10.6485051Z 2025-09-07T09:03:10.6485148Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6485595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6486024Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6486341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6486664Z x = module(x) 2025-09-07T09:03:10.6486945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T09:03:10.6487278Z out = self.conv2(out) 2025-09-07T09:03:10.6487386Z 2025-09-07T09:03:10.6487475Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6487927Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6488351Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6488664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6488989Z x = module(x) 2025-09-07T09:03:10.6489278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T09:03:10.6489652Z out = self.relu(out) 2025-09-07T09:03:10.6489744Z 2025-09-07T09:03:10.6489842Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6490285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6490752Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6491070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6491393Z x = module(x) 2025-09-07T09:03:10.6491674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T09:03:10.6492004Z out = self.conv3(out) 2025-09-07T09:03:10.6492108Z 2025-09-07T09:03:10.6492198Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6492646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6493078Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6493386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6493706Z x = module(x) 2025-09-07T09:03:10.6494033Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T09:03:10.6494366Z out = self.relu(out) 2025-09-07T09:03:10.6494457Z 2025-09-07T09:03:10.6494545Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6494997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6495457Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6495778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6496101Z x = module(x) 2025-09-07T09:03:10.6496383Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T09:03:10.6496716Z out = self.conv1(x) 2025-09-07T09:03:10.6496814Z 2025-09-07T09:03:10.6496902Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6497357Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6497871Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6498183Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6498510Z x = module(x) 2025-09-07T09:03:10.6498797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T09:03:10.6499130Z out = self.relu(out) 2025-09-07T09:03:10.6499219Z 2025-09-07T09:03:10.6499312Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6499758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6500185Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6500507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6500828Z x = module(x) 2025-09-07T09:03:10.6501108Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T09:03:10.6501440Z out = self.conv2(out) 2025-09-07T09:03:10.6501541Z 2025-09-07T09:03:10.6501631Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6502081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6502506Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6502858Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6503354Z x = module(x) 2025-09-07T09:03:10.6503642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T09:03:10.6504067Z out = self.relu(out) 2025-09-07T09:03:10.6504164Z 2025-09-07T09:03:10.6504257Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6504706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6505131Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6505451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6505773Z x = module(x) 2025-09-07T09:03:10.6506054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T09:03:10.6506385Z out = self.conv3(out) 2025-09-07T09:03:10.6506488Z 2025-09-07T09:03:10.6506578Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6507024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6507496Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6507813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6508136Z x = module(x) 2025-09-07T09:03:10.6508426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T09:03:10.6508758Z out = self.relu(out) 2025-09-07T09:03:10.6508898Z 2025-09-07T09:03:10.6508992Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6509439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6509868Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6510188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6510512Z x = module(x) 2025-09-07T09:03:10.6510798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T09:03:10.6511128Z out = self.conv1(x) 2025-09-07T09:03:10.6511218Z 2025-09-07T09:03:10.6511318Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6511770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6512192Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6512512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6512836Z x = module(x) 2025-09-07T09:03:10.6513123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T09:03:10.6513451Z out = self.relu(out) 2025-09-07T09:03:10.6513544Z 2025-09-07T09:03:10.6513634Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6514089Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6514515Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6514834Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6515152Z x = module(x) 2025-09-07T09:03:10.6515446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T09:03:10.6515776Z out = self.conv2(out) 2025-09-07T09:03:10.6515870Z 2025-09-07T09:03:10.6516022Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6516327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6516395Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6516599Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6516688Z x = module(x) 2025-09-07T09:03:10.6516899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T09:03:10.6516957Z out = self.relu(out) 2025-09-07T09:03:10.6516960Z 2025-09-07T09:03:10.6517060Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6517361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6517434Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6517638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6517697Z x = module(x) 2025-09-07T09:03:10.6517910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T09:03:10.6517970Z out = self.conv3(out) 2025-09-07T09:03:10.6517975Z 2025-09-07T09:03:10.6518111Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6518433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6518496Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6518755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6518809Z x = module(x) 2025-09-07T09:03:10.6519024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 158, in forward 2025-09-07T09:03:10.6519097Z identity = self.downsample(x) 2025-09-07T09:03:10.6519100Z 2025-09-07T09:03:10.6519198Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6519499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6519571Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6519778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6519829Z x = module(x) 2025-09-07T09:03:10.6520042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T09:03:10.6520101Z out = self.conv1(x) 2025-09-07T09:03:10.6520106Z 2025-09-07T09:03:10.6520195Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6520500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6520567Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6520776Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6520828Z x = module(x) 2025-09-07T09:03:10.6521034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T09:03:10.6521100Z out = self.relu(out) 2025-09-07T09:03:10.6521103Z 2025-09-07T09:03:10.6521191Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6521500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6521563Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6521769Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6521860Z x = module(x) 2025-09-07T09:03:10.6522067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T09:03:10.6522134Z out = self.conv2(out) 2025-09-07T09:03:10.6522136Z 2025-09-07T09:03:10.6522226Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6522569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6522632Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6522836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6522903Z x = module(x) 2025-09-07T09:03:10.6523107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T09:03:10.6523167Z out = self.relu(out) 2025-09-07T09:03:10.6523170Z 2025-09-07T09:03:10.6523261Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6523561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6523622Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6523860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6523915Z x = module(x) 2025-09-07T09:03:10.6524119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T09:03:10.6524177Z out = self.conv3(out) 2025-09-07T09:03:10.6524179Z 2025-09-07T09:03:10.6524265Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6524595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6524658Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6524877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6524938Z x = module(x) 2025-09-07T09:03:10.6525144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T09:03:10.6525209Z out = self.relu(out) 2025-09-07T09:03:10.6525213Z 2025-09-07T09:03:10.6525304Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6525602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6525671Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6525872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6525933Z x = module(x) 2025-09-07T09:03:10.6526136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 146, in forward 2025-09-07T09:03:10.6526194Z out = self.conv1(x) 2025-09-07T09:03:10.6526197Z 2025-09-07T09:03:10.6526293Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6526593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6526668Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6526869Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6526923Z x = module(x) 2025-09-07T09:03:10.6527136Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 148, in forward 2025-09-07T09:03:10.6527193Z out = self.relu(out) 2025-09-07T09:03:10.6527197Z 2025-09-07T09:03:10.6527298Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6527594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6527695Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6527897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6534206Z x = module(x) 2025-09-07T09:03:10.6534578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 150, in forward 2025-09-07T09:03:10.6534645Z out = self.conv2(out) 2025-09-07T09:03:10.6534649Z 2025-09-07T09:03:10.6534762Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6535095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6535172Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6535395Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6535452Z x = module(x) 2025-09-07T09:03:10.6535669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 152, in forward 2025-09-07T09:03:10.6535729Z out = self.relu(out) 2025-09-07T09:03:10.6535732Z 2025-09-07T09:03:10.6535834Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6536188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6536253Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6536461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6536512Z x = module(x) 2025-09-07T09:03:10.6536760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 154, in forward 2025-09-07T09:03:10.6536822Z out = self.conv3(out) 2025-09-07T09:03:10.6536825Z 2025-09-07T09:03:10.6536922Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6537224Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 190, in forward 2025-09-07T09:03:10.6537283Z feats = self.feat_extractor(x) 2025-09-07T09:03:10.6537491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/_utils.py", line 69, in forward 2025-09-07T09:03:10.6537545Z x = module(x) 2025-09-07T09:03:10.6537829Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torchvision/models/resnet.py", line 161, in forward 2025-09-07T09:03:10.6537887Z out = self.relu(out) 2025-09-07T09:03:10.6537891Z 2025-09-07T09:03:10.6537981Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6538285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6538351Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6538665Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T09:03:10.6538805Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T09:03:10.6539126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T09:03:10.6539259Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T09:03:10.6539262Z 2025-09-07T09:03:10.6539353Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6539674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6539740Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6540060Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T09:03:10.6540236Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T09:03:10.6540561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T09:03:10.6540726Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T09:03:10.6540729Z 2025-09-07T09:03:10.6540825Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6541138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6541200Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6541504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T09:03:10.6541630Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T09:03:10.6541935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T09:03:10.6542061Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T09:03:10.6542067Z 2025-09-07T09:03:10.6542190Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6542499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6542565Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6542902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 86, in forward 2025-09-07T09:03:10.6542980Z out.append(self.upsample(out[-1]) + t) 2025-09-07T09:03:10.6542983Z 2025-09-07T09:03:10.6543074Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6543375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6543435Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6543732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T09:03:10.6543854Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T09:03:10.6544161Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T09:03:10.6544286Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T09:03:10.6544289Z 2025-09-07T09:03:10.6544375Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6544675Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6544735Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6545032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 86, in forward 2025-09-07T09:03:10.6545104Z out.append(self.upsample(out[-1]) + t) 2025-09-07T09:03:10.6545107Z 2025-09-07T09:03:10.6545196Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6545491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6545551Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6545853Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in forward 2025-09-07T09:03:10.6545975Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T09:03:10.6546330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 83, in 2025-09-07T09:03:10.6546450Z _x: List[torch.Tensor] = [branch(t) for branch, t in zip(self.in_branches, x)] 2025-09-07T09:03:10.6546495Z 2025-09-07T09:03:10.6546588Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6546883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6546941Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6547243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 86, in forward 2025-09-07T09:03:10.6547311Z out.append(self.upsample(out[-1]) + t) 2025-09-07T09:03:10.6547314Z 2025-09-07T09:03:10.6547402Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6547702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6547763Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6548112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T09:03:10.6548252Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6548558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T09:03:10.6548664Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6548668Z 2025-09-07T09:03:10.6548796Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6549102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6549169Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6549467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T09:03:10.6549574Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6549882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T09:03:10.6549985Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6549989Z 2025-09-07T09:03:10.6550078Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6550371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6550433Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6550726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T09:03:10.6550830Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6551140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T09:03:10.6551246Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6551250Z 2025-09-07T09:03:10.6551341Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6551638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6551706Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6551997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T09:03:10.6552156Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6552464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T09:03:10.6552603Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6552607Z 2025-09-07T09:03:10.6552700Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6552993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6553056Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6553353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T09:03:10.6553459Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6553767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T09:03:10.6553872Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6553876Z 2025-09-07T09:03:10.6554001Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6554300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6554363Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6554655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T09:03:10.6554795Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6555112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T09:03:10.6555229Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6555233Z 2025-09-07T09:03:10.6555332Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6555634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6555705Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6556002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T09:03:10.6556115Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6556426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T09:03:10.6556532Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6556535Z 2025-09-07T09:03:10.6556630Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6556926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6556992Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6557286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in forward 2025-09-07T09:03:10.6557394Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6557699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 89, in 2025-09-07T09:03:10.6560414Z out = [branch(t) for branch, t in zip(self.out_branches, out[::-1])] 2025-09-07T09:03:10.6560420Z 2025-09-07T09:03:10.6560573Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6560880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 193, in forward 2025-09-07T09:03:10.6560951Z feat_concat = self.fpn(feats) 2025-09-07T09:03:10.6561254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 91, in forward 2025-09-07T09:03:10.6561364Z return torch.cat(out, dim=1) 2025-09-07T09:03:10.6561367Z 2025-09-07T09:03:10.6561443Z cudagraph partition due to non gpu ops 2025-09-07T09:03:10.6561534Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6561844Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T09:03:10.6561918Z logits = self.prob_head(feat_concat) 2025-09-07T09:03:10.6561921Z 2025-09-07T09:03:10.6562017Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6562318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T09:03:10.6562386Z logits = self.prob_head(feat_concat) 2025-09-07T09:03:10.6562392Z 2025-09-07T09:03:10.6562487Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6562827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T09:03:10.6562900Z logits = self.prob_head(feat_concat) 2025-09-07T09:03:10.6562903Z 2025-09-07T09:03:10.6562992Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6563327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 194, in forward 2025-09-07T09:03:10.6563393Z logits = self.prob_head(feat_concat) 2025-09-07T09:03:10.6563398Z 2025-09-07T09:03:10.6563469Z cudagraph partition due to non gpu ops 2025-09-07T09:03:10.6563564Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:10.6563860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/differentiable_binarization/pytorch.py", line 202, in forward 2025-09-07T09:03:10.6563953Z prob_map = _bf16_to_float32(torch.sigmoid(logits)) 2025-09-07T09:03:10.6564197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/utils/pytorch.py", line 30, in _bf16_to_float32 2025-09-07T09:03:10.6564298Z return x.float() if x.dtype == torch.bfloat16 else x 2025-09-07T09:03:10.6564302Z 2025-09-07T09:03:38.8860138Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:03:38.8860666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/doctr/models/detection/core.py", line 95, in __call__ 2025-09-07T09:03:38.8861117Z for bmap in (proba_map >= self.bin_thresh).astype(np.uint8) 2025-09-07T09:03:38.8861305Z 2025-09-07T09:03:39.9283827Z 2025-09-07T09:03:40.1450039Z running benchmark: 0% 0/30 [00:00blqk", from_blocked_mask[:, 2:-2], exp_blocked_to_pad) 2025-09-07T09:07:42.7872727Z 2025-09-07T09:07:44.1042150Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:44.1042716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1362, in forward 2025-09-07T09:07:44.1043164Z band_mask = band_mask.to(hidden_states.dtype) 2025-09-07T09:07:44.1043310Z 2025-09-07T09:07:44.1043417Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:44.1043902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1364, in forward 2025-09-07T09:07:44.1044330Z from_mask = from_mask.to(hidden_states.dtype) 2025-09-07T09:07:44.1044473Z 2025-09-07T09:07:44.1044582Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:44.1045020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1366, in forward 2025-09-07T09:07:44.1045425Z to_mask = to_mask.to(hidden_states.dtype) 2025-09-07T09:07:44.1045565Z 2025-09-07T09:07:44.8390165Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:44.8390717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 440, in forward 2025-09-07T09:07:44.8391149Z self.query(hidden_states) 2025-09-07T09:07:44.8391268Z 2025-09-07T09:07:44.8391406Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:44.8391878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 445, in forward 2025-09-07T09:07:44.8392288Z self.key(hidden_states) 2025-09-07T09:07:44.8392399Z 2025-09-07T09:07:44.8392508Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:44.8392946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 450, in forward 2025-09-07T09:07:44.8393340Z self.value(hidden_states) 2025-09-07T09:07:44.8393462Z 2025-09-07T09:07:44.9958512Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:44.9959120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1139, in _bigbird_block_rand_mask_with_head 2025-09-07T09:07:44.9959656Z plan_block_length = np.array(plan_from_length) // from_block_size 2025-09-07T09:07:44.9960187Z 2025-09-07T09:07:47.6163821Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6164811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 602, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6165471Z first_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, 0], key_layer, ndim=4) 2025-09-07T09:07:47.6166035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T09:07:47.6166511Z return torch.bmm( 2025-09-07T09:07:47.6166613Z 2025-09-07T09:07:47.6166696Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6166911Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6167102Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6167298Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6167485Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6167790Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6167988Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6168182Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6168375Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6168559Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6168752Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6169018Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6169219Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6169405Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6169596Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6169787Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6170008Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6170552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 647, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6171147Z to_mask.new_ones([bsz, 1, 1, n_rand_blocks * to_block_size]), 2025-09-07T09:07:47.6171327Z 2025-09-07T09:07:47.6171428Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6171985Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 653, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6172555Z rand_mask.new_ones([bsz, n_heads, from_block_size, 4 * to_block_size]), 2025-09-07T09:07:47.6172742Z 2025-09-07T09:07:47.6172819Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6173018Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6173236Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6173762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 698, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6174281Z first_band_product = torch.einsum( 2025-09-07T09:07:47.6174414Z 2025-09-07T09:07:47.6174485Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6174678Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6174873Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6175124Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6175666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 704, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6176173Z last_band_product = torch.einsum( 2025-09-07T09:07:47.6176300Z 2025-09-07T09:07:47.6176384Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6176580Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6176781Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6176976Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6177172Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6177429Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6177704Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6177931Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6178470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 781, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6179022Z to_mask.new_ones([bsz, 1, 1, n_rand_blocks * to_block_size]), 2025-09-07T09:07:47.6179189Z 2025-09-07T09:07:47.6179289Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6179824Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 787, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6180393Z rand_mask.new_ones([bsz, n_heads, from_block_size, 4 * to_block_size]), 2025-09-07T09:07:47.6180579Z 2025-09-07T09:07:47.6180704Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6180913Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6181131Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6181656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 807, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6183098Z last_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, -1], key_layer, ndim=4) 2025-09-07T09:07:47.6183661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T09:07:47.6184110Z return torch.bmm( 2025-09-07T09:07:47.6184211Z 2025-09-07T09:07:47.6184311Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6184879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 611, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6185482Z first_context_layer = self.torch_bmm_nd(first_attn_weights, value_layer, ndim=4) 2025-09-07T09:07:47.6185987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T09:07:47.6186519Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T09:07:47.6186754Z 2025-09-07T09:07:47.6186852Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6187390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 574, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6187904Z rand_attn = np.stack(rand_attn, axis=0) 2025-09-07T09:07:47.6188034Z 2025-09-07T09:07:47.6188134Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6188666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 575, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6189245Z rand_attn = torch.tensor(rand_attn, device=query_layer.device, dtype=torch.long) 2025-09-07T09:07:47.6189465Z 2025-09-07T09:07:47.6189561Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6190087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 620, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6190591Z second_key_mat = torch.cat( 2025-09-07T09:07:47.6190703Z 2025-09-07T09:07:47.6190800Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6191316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 642, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6192030Z second_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, 1], second_key_mat, ndim=4) 2025-09-07T09:07:47.6192622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T09:07:47.6193048Z return torch.bmm( 2025-09-07T09:07:47.6193155Z 2025-09-07T09:07:47.6193267Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6193806Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 643, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6194313Z second_seq_pad = torch.cat( 2025-09-07T09:07:47.6194439Z 2025-09-07T09:07:47.6194539Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6195122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 651, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6195642Z second_rand_pad = torch.cat( 2025-09-07T09:07:47.6195764Z 2025-09-07T09:07:47.6195865Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6196386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 630, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6196917Z second_value_mat = torch.cat( 2025-09-07T09:07:47.6197048Z 2025-09-07T09:07:47.6197141Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6197667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 665, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6198271Z second_context_layer = self.torch_bmm_nd(second_attn_weights, second_value_mat, ndim=4) 2025-09-07T09:07:47.6198778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T09:07:47.6199310Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T09:07:47.6199552Z 2025-09-07T09:07:47.6199646Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6200177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 676, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6200683Z exp_blocked_key_matrix = torch.cat( 2025-09-07T09:07:47.6200812Z 2025-09-07T09:07:47.6200909Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6201425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 687, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6202054Z inner_band_product = self.torch_bmm_nd_transpose(middle_query_matrix, exp_blocked_key_matrix, ndim=5) 2025-09-07T09:07:47.6202617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T09:07:47.6203230Z return torch.bmm( 2025-09-07T09:07:47.6203321Z 2025-09-07T09:07:47.6203419Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6203938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 693, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6204568Z rand_band_product = self.torch_bmm_nd_transpose(middle_query_matrix, gathered_key[:, :, 1:-1], ndim=5) 2025-09-07T09:07:47.6205123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T09:07:47.6205543Z return torch.bmm( 2025-09-07T09:07:47.6205695Z 2025-09-07T09:07:47.6205794Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6206362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 711, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6206955Z first_band_product += (1.0 - to_mask[:, :, :, :to_block_size].unsqueeze(3)) * attn_mask_penalty 2025-09-07T09:07:47.6207173Z 2025-09-07T09:07:47.6207268Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6207796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 710, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6208346Z inner_band_product += (1.0 - band_mask) * attn_mask_penalty 2025-09-07T09:07:47.6208512Z 2025-09-07T09:07:47.6208610Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6209168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 713, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6209740Z rand_band_product += (1.0 - rand_mask[:, :, 1:-1]) * attn_mask_penalty 2025-09-07T09:07:47.6209929Z 2025-09-07T09:07:47.6210029Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6210620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 712, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6211220Z last_band_product += (1.0 - to_mask[:, :, :, -to_block_size:].unsqueeze(3)) * attn_mask_penalty 2025-09-07T09:07:47.6211431Z 2025-09-07T09:07:47.6211537Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6212071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 716, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6212576Z band_product = torch.cat( 2025-09-07T09:07:47.6212693Z 2025-09-07T09:07:47.6212797Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6213325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 679, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6213833Z exp_blocked_value_matrix = torch.cat( 2025-09-07T09:07:47.6213973Z 2025-09-07T09:07:47.6214063Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6214588Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 727, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6215100Z context_layer = self.torch_bmm_nd( 2025-09-07T09:07:47.6215511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T09:07:47.6216039Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T09:07:47.6216284Z 2025-09-07T09:07:47.6216375Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6216903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 734, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6217426Z context_layer += self.torch_bmm_nd( 2025-09-07T09:07:47.6217872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T09:07:47.6218391Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T09:07:47.6218623Z 2025-09-07T09:07:47.6218716Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6219241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 740, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6219825Z context_layer += torch.einsum( 2025-09-07T09:07:47.6219948Z 2025-09-07T09:07:47.6220047Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6220571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 743, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6221070Z context_layer += torch.einsum( 2025-09-07T09:07:47.6221191Z 2025-09-07T09:07:47.6221282Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6221805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 754, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6222307Z second_last_key_mat = torch.cat( 2025-09-07T09:07:47.6222432Z 2025-09-07T09:07:47.6222565Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6223081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 776, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6223726Z second_last_product = self.torch_bmm_nd_transpose(blocked_query_matrix[:, :, -2], second_last_key_mat, ndim=4) 2025-09-07T09:07:47.6224331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 493, in torch_bmm_nd_transpose 2025-09-07T09:07:47.6224761Z return torch.bmm( 2025-09-07T09:07:47.6224849Z 2025-09-07T09:07:47.6224944Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6225462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 777, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6225970Z second_last_seq_pad = torch.cat( 2025-09-07T09:07:47.6226097Z 2025-09-07T09:07:47.6226191Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6226711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 785, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6227214Z second_last_rand_pad = torch.cat( 2025-09-07T09:07:47.6227338Z 2025-09-07T09:07:47.6227440Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6227958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 764, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6228461Z second_last_value_mat = torch.cat( 2025-09-07T09:07:47.6228592Z 2025-09-07T09:07:47.6228683Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6229207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 799, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6229835Z second_last_context_layer = self.torch_bmm_nd(second_last_attn_weights, second_last_value_mat, ndim=4) 2025-09-07T09:07:47.6230360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T09:07:47.6230879Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T09:07:47.6231116Z 2025-09-07T09:07:47.6231206Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6231734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 813, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6232311Z last_context_layer = self.torch_bmm_nd(last_attn_weights, value_layer, ndim=4) 2025-09-07T09:07:47.6232900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 485, in torch_bmm_nd 2025-09-07T09:07:47.6233452Z return torch.bmm(inp_1.reshape((-1,) + inp_1.shape[-2:]), inp_2.reshape((-1,) + inp_2.shape[-2:])).view( 2025-09-07T09:07:47.6233686Z 2025-09-07T09:07:47.6233761Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6233964Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6234199Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6234713Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 743, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6235216Z context_layer += torch.einsum( 2025-09-07T09:07:47.6235350Z 2025-09-07T09:07:47.6235420Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6235630Z cudagraph partition due to non gpu ops 2025-09-07T09:07:47.6235883Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6236403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 817, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6236911Z context_layer = torch.cat( 2025-09-07T09:07:47.6237026Z 2025-09-07T09:07:47.6237116Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:47.6237686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 821, in torch_dynamo_resume_in_bigbird_block_sparse_attention_at_564 2025-09-07T09:07:47.6238280Z context_layer = context_layer.view((bsz, n_heads, from_seq_len, -1)) * from_mask 2025-09-07T09:07:47.6238492Z 2025-09-07T09:07:56.4275723Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:56.4276365Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 478, in torch_dynamo_resume_in_forward_at_455 2025-09-07T09:07:56.4276967Z context_layer = context_layer.contiguous().view(batch_size, from_seq_length, -1) 2025-09-07T09:07:56.4277196Z 2025-09-07T09:07:57.1445857Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:57.1446477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1385, in torch_dynamo_resume_in_forward_at_1381 2025-09-07T09:07:57.1447019Z attention_output = self.output(self_outputs[0], hidden_states) 2025-09-07T09:07:57.1447506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1295, in forward 2025-09-07T09:07:57.1447934Z hidden_states = self.dense(hidden_states) 2025-09-07T09:07:57.1448074Z 2025-09-07T09:07:58.1111306Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:58.1111904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T09:07:58.1112408Z layer_output = apply_chunking_to_forward( 2025-09-07T09:07:58.1112828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:07:58.1113241Z return forward_fn(*input_tensors) 2025-09-07T09:07:58.1113711Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T09:07:58.1114197Z intermediate_output = self.intermediate(attention_output) 2025-09-07T09:07:58.1114641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1401, in forward 2025-09-07T09:07:58.1115058Z hidden_states = self.dense(hidden_states) 2025-09-07T09:07:58.1115200Z 2025-09-07T09:07:58.1115302Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:58.1116108Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T09:07:58.1116693Z layer_output = apply_chunking_to_forward( 2025-09-07T09:07:58.1117093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:07:58.1117490Z return forward_fn(*input_tensors) 2025-09-07T09:07:58.1117913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1512, in feed_forward_chunk 2025-09-07T09:07:58.1118384Z intermediate_output = self.intermediate(attention_output) 2025-09-07T09:07:58.1118822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1402, in forward 2025-09-07T09:07:58.1119252Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T09:07:58.1119719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 47, in forward 2025-09-07T09:07:58.1120191Z return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi) * (input + 0.044715 * torch.pow(input, 3.0)))) 2025-09-07T09:07:58.1120431Z 2025-09-07T09:07:58.1120542Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:07:58.1121109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1505, in torch_dynamo_resume_in_forward_at_1468 2025-09-07T09:07:58.1121585Z layer_output = apply_chunking_to_forward( 2025-09-07T09:07:58.1121988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:07:58.1122383Z return forward_fn(*input_tensors) 2025-09-07T09:07:58.1122799Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1513, in feed_forward_chunk 2025-09-07T09:07:58.1123293Z layer_output = self.output(intermediate_output, attention_output) 2025-09-07T09:07:58.1123750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1415, in forward 2025-09-07T09:07:58.1124162Z hidden_states = self.dense(hidden_states) 2025-09-07T09:07:58.1124307Z 2025-09-07T09:08:05.3333183Z W0907 09:08:05.332554 168802 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] torch._dynamo hit config.recompile_limit (8) 2025-09-07T09:08:05.3334104Z W0907 09:08:05.332554 168802 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] function: 'forward' (/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py:417) 2025-09-07T09:08:05.3335870Z W0907 09:08:05.332554 168802 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] last reason: 10/7: self.seed == 7 # context_layer, attention_probs = self.bigbird_block_sparse_attention( # transformers/models/big_bird/modeling_big_bird.py:455 in forward (HINT: torch.compile considers integer attributes of the nn.Module to be static. If you are observing recompilation, you might want to make this integer dynamic using torch._dynamo.config.allow_unspec_int_on_nn_module = True, or convert this integer into a tensor.) 2025-09-07T09:08:05.3337470Z W0907 09:08:05.332554 168802 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] To log all recompilation reasons, use TORCH_LOGS="recompiles". 2025-09-07T09:08:05.3338355Z W0907 09:08:05.332554 168802 site-packages/torch/_dynamo/convert_frame.py:1358] [10/8] To diagnose recompilation issues, see https://pytorch.org/docs/main/torch.compiler_troubleshooting.html 2025-09-07T09:08:05.9734315Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:08:05.9734948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2010, in torch_dynamo_resume_in_forward_at_1991 2025-09-07T09:08:05.9735990Z pooler_output = self.activation(self.pooler(sequence_output[:, 0, :])) if (self.pooler is not None) else None 2025-09-07T09:08:05.9736364Z 2025-09-07T09:08:05.9736481Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:08:05.9736983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2010, in torch_dynamo_resume_in_forward_at_1991 2025-09-07T09:08:05.9737663Z pooler_output = self.activation(self.pooler(sequence_output[:, 0, :])) if (self.pooler is not None) else None 2025-09-07T09:08:05.9737937Z 2025-09-07T09:08:06.1779177Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:08:06.1779788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2318, in torch_dynamo_resume_in_forward_at_2303 2025-09-07T09:08:06.1780301Z prediction_scores = self.cls(sequence_output) 2025-09-07T09:08:06.1781121Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1682, in forward 2025-09-07T09:08:06.1781587Z prediction_scores = self.predictions(sequence_output) 2025-09-07T09:08:06.1782035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1670, in forward 2025-09-07T09:08:06.1782458Z hidden_states = self.transform(hidden_states) 2025-09-07T09:08:06.1782987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1645, in forward 2025-09-07T09:08:06.1783403Z hidden_states = self.dense(hidden_states) 2025-09-07T09:08:06.1783541Z 2025-09-07T09:08:06.1783648Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:08:06.1784152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 2318, in torch_dynamo_resume_in_forward_at_2303 2025-09-07T09:08:06.1784646Z prediction_scores = self.cls(sequence_output) 2025-09-07T09:08:06.1785058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1682, in forward 2025-09-07T09:08:06.1785497Z prediction_scores = self.predictions(sequence_output) 2025-09-07T09:08:06.1785915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/big_bird/modeling_big_bird.py", line 1671, in forward 2025-09-07T09:08:06.1786332Z hidden_states = self.decoder(hidden_states) 2025-09-07T09:08:06.1786476Z 2025-09-07T09:08:08.8335327Z 2025-09-07T09:08:09.8013254Z running benchmark: 0% 0/30 [00:00bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1358602Z 2025-09-07T09:11:22.1358713Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1359225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1359707Z layer_outputs = layer_module( 2025-09-07T09:11:22.1360080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1360457Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1360863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1361284Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1361694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1362104Z self_outputs = self.self( 2025-09-07T09:11:22.1362497Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1362934Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1363435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1364023Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1364262Z 2025-09-07T09:11:22.1364350Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1364555Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1364755Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1364948Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1365174Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1365686Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1366163Z layer_outputs = layer_module( 2025-09-07T09:11:22.1366500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1366857Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1367277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1367698Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1368110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1368592Z self_outputs = self.self( 2025-09-07T09:11:22.1368986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T09:11:22.1369442Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1369969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1370528Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T09:11:22.1371065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T09:11:22.1371600Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T09:11:22.1371805Z 2025-09-07T09:11:22.1371893Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1372170Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1372696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1373202Z layer_outputs = layer_module( 2025-09-07T09:11:22.1373549Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1373909Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1374373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1374796Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1375207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1375638Z self_outputs = self.self( 2025-09-07T09:11:22.1376050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:11:22.1376472Z attn_scores += diagonal_mask 2025-09-07T09:11:22.1376597Z 2025-09-07T09:11:22.1376705Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1377209Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1377695Z layer_outputs = layer_module( 2025-09-07T09:11:22.1378112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1378475Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1378892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1379302Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1379712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1380126Z self_outputs = self.self( 2025-09-07T09:11:22.1380523Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:11:22.1380950Z attn_probs = nn.functional.softmax( 2025-09-07T09:11:22.1381072Z 2025-09-07T09:11:22.1381164Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1381696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1382187Z layer_outputs = layer_module( 2025-09-07T09:11:22.1382525Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1382938Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1383382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1383794Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1384207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1384633Z self_outputs = self.self( 2025-09-07T09:11:22.1385016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1385477Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1386019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1386653Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:11:22.1387092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1387436Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1387594Z 2025-09-07T09:11:22.1387690Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1388206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1388693Z layer_outputs = layer_module( 2025-09-07T09:11:22.1389087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1389450Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1389861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1390280Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1390691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1391098Z self_outputs = self.self( 2025-09-07T09:11:22.1391487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1391940Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1392471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1393026Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:11:22.1393548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:11:22.1394011Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:11:22.1394348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1394694Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1394842Z 2025-09-07T09:11:22.1394950Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1395478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1395970Z layer_outputs = layer_module( 2025-09-07T09:11:22.1396321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1396682Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1397100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1397607Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1398042Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1398469Z self_outputs = self.self( 2025-09-07T09:11:22.1398870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1399333Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1399867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1400435Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1400649Z 2025-09-07T09:11:22.1400750Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1401311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1401806Z layer_outputs = layer_module( 2025-09-07T09:11:22.1402141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1402491Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1402909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1403583Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1404004Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1404419Z self_outputs = self.self( 2025-09-07T09:11:22.1404818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1405281Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1405815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1406379Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1406584Z 2025-09-07T09:11:22.1406693Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1407206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1407687Z layer_outputs = layer_module( 2025-09-07T09:11:22.1408019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1408380Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1408788Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1409208Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1409618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1410025Z self_outputs = self.self( 2025-09-07T09:11:22.1410420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:11:22.1410940Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:11:22.1411184Z 2025-09-07T09:11:22.1411277Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1411790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1412396Z layer_outputs = layer_module( 2025-09-07T09:11:22.1412730Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1413078Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1413491Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1413907Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1414319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T09:11:22.1414766Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T09:11:22.1415207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T09:11:22.1415676Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1415821Z 2025-09-07T09:11:22.1415916Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1416430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1416915Z layer_outputs = layer_module( 2025-09-07T09:11:22.1417244Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1417598Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1418105Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1418536Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1418945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1419344Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1419763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1420226Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1420673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T09:11:22.1421095Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1421229Z 2025-09-07T09:11:22.1421329Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1421841Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1422329Z layer_outputs = layer_module( 2025-09-07T09:11:22.1422672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1423026Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1423437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1423860Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1424258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1424651Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1425058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1425508Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1425956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T09:11:22.1426488Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T09:11:22.1426863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T09:11:22.1427204Z return self.act(input) 2025-09-07T09:11:22.1427317Z 2025-09-07T09:11:22.1427413Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1427923Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1428416Z layer_outputs = layer_module( 2025-09-07T09:11:22.1428753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1429103Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1429553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1429982Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1430384Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1430779Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1431186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T09:11:22.1431653Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T09:11:22.1432153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T09:11:22.1432594Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1432732Z 2025-09-07T09:11:22.1432836Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1433360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1433853Z layer_outputs = layer_module( 2025-09-07T09:11:22.1434203Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1434565Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1434975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1435401Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1435820Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1436231Z self_outputs = self.self( 2025-09-07T09:11:22.1436639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:11:22.1437054Z query_vectors = self.query(hidden_states) 2025-09-07T09:11:22.1437198Z 2025-09-07T09:11:22.1437294Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1437805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1438295Z layer_outputs = layer_module( 2025-09-07T09:11:22.1438638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1438991Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1439404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1439824Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1440235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1440723Z self_outputs = self.self( 2025-09-07T09:11:22.1441113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1441554Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1442061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1442668Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1442920Z 2025-09-07T09:11:22.1443023Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1443528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1444062Z layer_outputs = layer_module( 2025-09-07T09:11:22.1444402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1444761Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1455250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1455718Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1456182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1456695Z self_outputs = self.self( 2025-09-07T09:11:22.1457125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T09:11:22.1457547Z key_vectors = self.key(hidden_states) 2025-09-07T09:11:22.1457691Z 2025-09-07T09:11:22.1457894Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1458429Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1458936Z layer_outputs = layer_module( 2025-09-07T09:11:22.1459285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1459658Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1460093Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1460525Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1460947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1461355Z self_outputs = self.self( 2025-09-07T09:11:22.1461756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1462211Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1462724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1463327Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1463574Z 2025-09-07T09:11:22.1463656Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1463873Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1464076Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1464274Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1464501Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1465029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1465609Z layer_outputs = layer_module( 2025-09-07T09:11:22.1465964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1466330Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1466744Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1467169Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1467592Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1468006Z self_outputs = self.self( 2025-09-07T09:11:22.1468421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T09:11:22.1468938Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1469464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1470021Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T09:11:22.1470562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T09:11:22.1471102Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T09:11:22.1471346Z 2025-09-07T09:11:22.1471426Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1471666Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1472193Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1472695Z layer_outputs = layer_module( 2025-09-07T09:11:22.1473048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1473412Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1473835Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1474261Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1474678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1475091Z self_outputs = self.self( 2025-09-07T09:11:22.1475484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:11:22.1475902Z attn_scores += diagonal_mask 2025-09-07T09:11:22.1476027Z 2025-09-07T09:11:22.1476350Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1476873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1477368Z layer_outputs = layer_module( 2025-09-07T09:11:22.1477703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1478081Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1478501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1478924Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1479325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1479735Z self_outputs = self.self( 2025-09-07T09:11:22.1480174Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:11:22.1480652Z attn_probs = nn.functional.softmax( 2025-09-07T09:11:22.1480780Z 2025-09-07T09:11:22.1480888Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1481396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1481887Z layer_outputs = layer_module( 2025-09-07T09:11:22.1482225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1482585Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1483000Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1483414Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1483857Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1484277Z self_outputs = self.self( 2025-09-07T09:11:22.1484672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T09:11:22.1485101Z value_vectors = self.value(hidden_states) 2025-09-07T09:11:22.1485238Z 2025-09-07T09:11:22.1485336Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1485884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1486374Z layer_outputs = layer_module( 2025-09-07T09:11:22.1486715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1487077Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1487486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1487904Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1488312Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1488720Z self_outputs = self.self( 2025-09-07T09:11:22.1489110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1489569Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1490098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1490690Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:11:22.1491116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1491461Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1491609Z 2025-09-07T09:11:22.1491714Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1492219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1492709Z layer_outputs = layer_module( 2025-09-07T09:11:22.1493049Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1493408Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1493815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1494299Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1494750Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1495162Z self_outputs = self.self( 2025-09-07T09:11:22.1495558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1496008Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1496538Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1497087Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:11:22.1497596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:11:22.1498167Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:11:22.1498498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1498845Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1499001Z 2025-09-07T09:11:22.1499096Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1499612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1500099Z layer_outputs = layer_module( 2025-09-07T09:11:22.1500475Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1500842Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1501260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1501680Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1502091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1502496Z self_outputs = self.self( 2025-09-07T09:11:22.1502893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1503525Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1504056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1504626Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1504832Z 2025-09-07T09:11:22.1504931Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1505452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1505957Z layer_outputs = layer_module( 2025-09-07T09:11:22.1506292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1506649Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1507059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1507476Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1507884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1508292Z self_outputs = self.self( 2025-09-07T09:11:22.1508674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1509289Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1509812Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1510374Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1510578Z 2025-09-07T09:11:22.1510684Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1511196Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1511677Z layer_outputs = layer_module( 2025-09-07T09:11:22.1512015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1512378Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1512876Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1513301Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1513720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1514133Z self_outputs = self.self( 2025-09-07T09:11:22.1514533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:11:22.1515114Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:11:22.1515359Z 2025-09-07T09:11:22.1515463Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1515982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1516475Z layer_outputs = layer_module( 2025-09-07T09:11:22.1516825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1517191Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1517603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1518019Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1518433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T09:11:22.1518883Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T09:11:22.1519328Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T09:11:22.1519758Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1519902Z 2025-09-07T09:11:22.1520000Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1520512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1521004Z layer_outputs = layer_module( 2025-09-07T09:11:22.1521343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1521701Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1522116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1522535Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1522935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1523385Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1523837Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1524294Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1524733Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T09:11:22.1525156Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1525292Z 2025-09-07T09:11:22.1525394Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1525905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1526391Z layer_outputs = layer_module( 2025-09-07T09:11:22.1526753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1527122Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1527539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1527966Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1528358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1528758Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1529207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1529667Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1530108Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T09:11:22.1530555Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T09:11:22.1530936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T09:11:22.1531281Z return self.act(input) 2025-09-07T09:11:22.1531389Z 2025-09-07T09:11:22.1531493Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1532006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1532486Z layer_outputs = layer_module( 2025-09-07T09:11:22.1532827Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1533184Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1533606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1534026Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1534420Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1534814Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1535236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T09:11:22.1535701Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T09:11:22.1536155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T09:11:22.1536581Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1536719Z 2025-09-07T09:11:22.1536816Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1537330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1539253Z layer_outputs = layer_module( 2025-09-07T09:11:22.1539587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1539946Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1540355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1540773Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1541187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1541605Z self_outputs = self.self( 2025-09-07T09:11:22.1541999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:11:22.1542463Z query_vectors = self.query(hidden_states) 2025-09-07T09:11:22.1542599Z 2025-09-07T09:11:22.1542703Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1543213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1543694Z layer_outputs = layer_module( 2025-09-07T09:11:22.1544034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1544390Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1544839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1545264Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1545669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1546079Z self_outputs = self.self( 2025-09-07T09:11:22.1546478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1546921Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1547425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1548021Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1548273Z 2025-09-07T09:11:22.1548370Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1548882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1549370Z layer_outputs = layer_module( 2025-09-07T09:11:22.1549710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1550078Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1550498Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1550917Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1551347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1551762Z self_outputs = self.self( 2025-09-07T09:11:22.1552168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T09:11:22.1552592Z key_vectors = self.key(hidden_states) 2025-09-07T09:11:22.1552722Z 2025-09-07T09:11:22.1552829Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1553397Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1553919Z layer_outputs = layer_module( 2025-09-07T09:11:22.1554268Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1554632Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1555055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1555480Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1555887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1556314Z self_outputs = self.self( 2025-09-07T09:11:22.1556756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1557224Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1557740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1558328Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1558576Z 2025-09-07T09:11:22.1558650Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1558844Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1559065Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1559258Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1559478Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1559989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1560473Z layer_outputs = layer_module( 2025-09-07T09:11:22.1560805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1561156Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1561564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1561968Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1562368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1562764Z self_outputs = self.self( 2025-09-07T09:11:22.1563140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T09:11:22.1563586Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1564084Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1564621Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T09:11:22.1565140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T09:11:22.1565671Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T09:11:22.1565893Z 2025-09-07T09:11:22.1565981Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1566245Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1566785Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1567345Z layer_outputs = layer_module( 2025-09-07T09:11:22.1567676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1568084Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1568511Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1568924Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1569331Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1569732Z self_outputs = self.self( 2025-09-07T09:11:22.1570125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:11:22.1570534Z attn_scores += diagonal_mask 2025-09-07T09:11:22.1570646Z 2025-09-07T09:11:22.1570781Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1571293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1571784Z layer_outputs = layer_module( 2025-09-07T09:11:22.1572107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1572483Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1572886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1573323Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1573725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1574125Z self_outputs = self.self( 2025-09-07T09:11:22.1574507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:11:22.1574913Z attn_probs = nn.functional.softmax( 2025-09-07T09:11:22.1575048Z 2025-09-07T09:11:22.1575138Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1575646Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1576143Z layer_outputs = layer_module( 2025-09-07T09:11:22.1576487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1576837Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1577235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1577637Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1578081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1578481Z self_outputs = self.self( 2025-09-07T09:11:22.1578862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T09:11:22.1579286Z value_vectors = self.value(hidden_states) 2025-09-07T09:11:22.1579431Z 2025-09-07T09:11:22.1579521Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1580053Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1580528Z layer_outputs = layer_module( 2025-09-07T09:11:22.1580850Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1581248Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1581676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1582125Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1582530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1582927Z self_outputs = self.self( 2025-09-07T09:11:22.1583310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1583759Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1584299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1584930Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:11:22.1585351Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1585704Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1585859Z 2025-09-07T09:11:22.1585954Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1586460Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1586938Z layer_outputs = layer_module( 2025-09-07T09:11:22.1587310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1587662Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1588074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1588502Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1588899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1589299Z self_outputs = self.self( 2025-09-07T09:11:22.1589710Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1590158Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1590683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1591224Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:11:22.1591725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:11:22.1592200Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:11:22.1592528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1592865Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1593010Z 2025-09-07T09:11:22.1593115Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1593621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1594121Z layer_outputs = layer_module( 2025-09-07T09:11:22.1594453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1594803Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1595211Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1595756Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1596199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1596604Z self_outputs = self.self( 2025-09-07T09:11:22.1596990Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1597434Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1597964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1598526Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1598730Z 2025-09-07T09:11:22.1598834Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1599393Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1599880Z layer_outputs = layer_module( 2025-09-07T09:11:22.1600218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1600571Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1600979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1601395Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1601830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1602247Z self_outputs = self.self( 2025-09-07T09:11:22.1602637Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1603215Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1603740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1604298Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1604506Z 2025-09-07T09:11:22.1604602Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1605112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1605596Z layer_outputs = layer_module( 2025-09-07T09:11:22.1605934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1606286Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1606698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1607112Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1607518Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1607920Z self_outputs = self.self( 2025-09-07T09:11:22.1608310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:11:22.1608835Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:11:22.1609073Z 2025-09-07T09:11:22.1609176Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1609690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1610281Z layer_outputs = layer_module( 2025-09-07T09:11:22.1610671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1611031Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1611441Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1611855Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1612262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T09:11:22.1612710Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T09:11:22.1613153Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T09:11:22.1613581Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1613773Z 2025-09-07T09:11:22.1613878Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1614385Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1614868Z layer_outputs = layer_module( 2025-09-07T09:11:22.1615200Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1615556Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1616035Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1616459Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1616861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1617260Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1617680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1618188Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1618633Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T09:11:22.1619051Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1619195Z 2025-09-07T09:11:22.1619293Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1619805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1620298Z layer_outputs = layer_module( 2025-09-07T09:11:22.1620627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1620985Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1621394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1621817Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1622207Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1622596Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1623008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1623458Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1623896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T09:11:22.1624400Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T09:11:22.1624816Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T09:11:22.1625156Z return self.act(input) 2025-09-07T09:11:22.1625263Z 2025-09-07T09:11:22.1625365Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1625873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1626357Z layer_outputs = layer_module( 2025-09-07T09:11:22.1626697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1627050Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1627460Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1627922Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1628316Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1628708Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1629116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T09:11:22.1629579Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T09:11:22.1630064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T09:11:22.1630489Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1630627Z 2025-09-07T09:11:22.1630722Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1631229Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1631716Z layer_outputs = layer_module( 2025-09-07T09:11:22.1632043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1632394Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1632808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1633227Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1633639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1634040Z self_outputs = self.self( 2025-09-07T09:11:22.1634433Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:11:22.1634856Z query_vectors = self.query(hidden_states) 2025-09-07T09:11:22.1634989Z 2025-09-07T09:11:22.1635089Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1635595Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1636074Z layer_outputs = layer_module( 2025-09-07T09:11:22.1636405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1636758Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1637172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1637586Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1637989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1638453Z self_outputs = self.self( 2025-09-07T09:11:22.1638877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1639317Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1639809Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1640409Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1640658Z 2025-09-07T09:11:22.1640754Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1641256Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1641743Z layer_outputs = layer_module( 2025-09-07T09:11:22.1642109Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1642464Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1642877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1643290Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1643701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1644107Z self_outputs = self.self( 2025-09-07T09:11:22.1644528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T09:11:22.1644949Z key_vectors = self.key(hidden_states) 2025-09-07T09:11:22.1645078Z 2025-09-07T09:11:22.1645182Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1645703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1646184Z layer_outputs = layer_module( 2025-09-07T09:11:22.1646515Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1646869Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1647279Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1647695Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1648097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1648505Z self_outputs = self.self( 2025-09-07T09:11:22.1648893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1649333Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1649830Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1650408Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1650652Z 2025-09-07T09:11:22.1650729Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1650932Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1651133Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1651322Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1651544Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1652061Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1652600Z layer_outputs = layer_module( 2025-09-07T09:11:22.1652968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1653319Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1653729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1654143Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1654553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1654963Z self_outputs = self.self( 2025-09-07T09:11:22.1655343Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T09:11:22.1655792Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1656341Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1656899Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T09:11:22.1657432Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T09:11:22.1658016Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T09:11:22.1658227Z 2025-09-07T09:11:22.1658341Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1658572Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1659087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1659578Z layer_outputs = layer_module( 2025-09-07T09:11:22.1659909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1660270Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1660681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1661094Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1661493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1661906Z self_outputs = self.self( 2025-09-07T09:11:22.1662294Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:11:22.1662705Z attn_scores += diagonal_mask 2025-09-07T09:11:22.1662820Z 2025-09-07T09:11:22.1662920Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1663421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1663906Z layer_outputs = layer_module( 2025-09-07T09:11:22.1664239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1664593Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1665001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1665409Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1665812Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1666224Z self_outputs = self.self( 2025-09-07T09:11:22.1666617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:11:22.1667104Z attn_probs = nn.functional.softmax( 2025-09-07T09:11:22.1667227Z 2025-09-07T09:11:22.1667320Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1667829Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1668309Z layer_outputs = layer_module( 2025-09-07T09:11:22.1668638Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1668993Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1669403Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1669816Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1670276Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1670691Z self_outputs = self.self( 2025-09-07T09:11:22.1671075Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T09:11:22.1671500Z value_vectors = self.value(hidden_states) 2025-09-07T09:11:22.1671640Z 2025-09-07T09:11:22.1671731Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1672274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1672764Z layer_outputs = layer_module( 2025-09-07T09:11:22.1673088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1673446Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1673859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1674275Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1674680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1675083Z self_outputs = self.self( 2025-09-07T09:11:22.1675477Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1675934Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1676462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1677047Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:11:22.1677462Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1677811Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1677964Z 2025-09-07T09:11:22.1678061Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1678573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1679055Z layer_outputs = layer_module( 2025-09-07T09:11:22.1679390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1679745Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1680154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1680620Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1681029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1681474Z self_outputs = self.self( 2025-09-07T09:11:22.1681864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1682322Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1682851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1683392Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:11:22.1683904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:11:22.1684373Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:11:22.1684748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1685110Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1685263Z 2025-09-07T09:11:22.1685367Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1685893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1686384Z layer_outputs = layer_module( 2025-09-07T09:11:22.1686773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1687149Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1687568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1687998Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1688418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1688842Z self_outputs = self.self( 2025-09-07T09:11:22.1689238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1689696Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1690236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1690805Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1691009Z 2025-09-07T09:11:22.1691116Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1691634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1692121Z layer_outputs = layer_module( 2025-09-07T09:11:22.1692458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1692818Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1693238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1693658Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1694069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1694481Z self_outputs = self.self( 2025-09-07T09:11:22.1694881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1695408Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1695963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1696526Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1696731Z 2025-09-07T09:11:22.1696831Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1697342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1697892Z layer_outputs = layer_module( 2025-09-07T09:11:22.1698230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1698583Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1699056Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1699485Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1699899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1700306Z self_outputs = self.self( 2025-09-07T09:11:22.1700700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:11:22.1701263Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:11:22.1701506Z 2025-09-07T09:11:22.1701611Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1702120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1702605Z layer_outputs = layer_module( 2025-09-07T09:11:22.1702938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1703431Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1703856Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1704276Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1704690Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T09:11:22.1705152Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T09:11:22.1705600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T09:11:22.1706025Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1706164Z 2025-09-07T09:11:22.1706273Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1706780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1707268Z layer_outputs = layer_module( 2025-09-07T09:11:22.1707611Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1707970Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1708419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1708845Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1709251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1709738Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1710160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1710671Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1711111Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T09:11:22.1711539Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1711677Z 2025-09-07T09:11:22.1711778Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1712298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1712792Z layer_outputs = layer_module( 2025-09-07T09:11:22.1713128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1713539Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1713957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1714383Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1714773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1715171Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1715640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1716107Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1716550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T09:11:22.1716999Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T09:11:22.1717381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T09:11:22.1717727Z return self.act(input) 2025-09-07T09:11:22.1717834Z 2025-09-07T09:11:22.1717939Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1718454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1718934Z layer_outputs = layer_module( 2025-09-07T09:11:22.1719274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1719633Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1720052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1720488Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1720888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1721284Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1721701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T09:11:22.1722170Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T09:11:22.1722625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T09:11:22.1723051Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1723192Z 2025-09-07T09:11:22.1723290Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1723801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1724390Z layer_outputs = layer_module( 2025-09-07T09:11:22.1724732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1725093Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1725506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1725928Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1726342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1726748Z self_outputs = self.self( 2025-09-07T09:11:22.1727143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:11:22.1727566Z query_vectors = self.query(hidden_states) 2025-09-07T09:11:22.1727733Z 2025-09-07T09:11:22.1727839Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1728348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1728823Z layer_outputs = layer_module( 2025-09-07T09:11:22.1729154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1729507Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1729945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1730366Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1730768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1731183Z self_outputs = self.self( 2025-09-07T09:11:22.1731576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1732018Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1732512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1733112Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1733364Z 2025-09-07T09:11:22.1733465Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1733982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1734468Z layer_outputs = layer_module( 2025-09-07T09:11:22.1734804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1735161Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1735575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1735989Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1736402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1736803Z self_outputs = self.self( 2025-09-07T09:11:22.1737195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T09:11:22.1737613Z key_vectors = self.key(hidden_states) 2025-09-07T09:11:22.1737785Z 2025-09-07T09:11:22.1737889Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1738439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1738958Z layer_outputs = layer_module( 2025-09-07T09:11:22.1739291Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1739643Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1740052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1740462Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1740863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1741269Z self_outputs = self.self( 2025-09-07T09:11:22.1741706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1742158Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1742659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1743237Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1743488Z 2025-09-07T09:11:22.1743565Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1743769Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1744004Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1744197Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1744420Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1744931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1745422Z layer_outputs = layer_module( 2025-09-07T09:11:22.1745757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1746109Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1746519Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1746933Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1747340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1747745Z self_outputs = self.self( 2025-09-07T09:11:22.1748127Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T09:11:22.1748576Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1749088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1749639Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T09:11:22.1750186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T09:11:22.1750713Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T09:11:22.1750922Z 2025-09-07T09:11:22.1750998Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1751226Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1751737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1752223Z layer_outputs = layer_module( 2025-09-07T09:11:22.1752602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1752998Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1753412Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1753830Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1754232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1754641Z self_outputs = self.self( 2025-09-07T09:11:22.1755034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:11:22.1755445Z attn_scores += diagonal_mask 2025-09-07T09:11:22.1755561Z 2025-09-07T09:11:22.1755664Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1756206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1756699Z layer_outputs = layer_module( 2025-09-07T09:11:22.1757040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1757395Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1757807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1758248Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1758663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1759071Z self_outputs = self.self( 2025-09-07T09:11:22.1759463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:11:22.1759884Z attn_probs = nn.functional.softmax( 2025-09-07T09:11:22.1760008Z 2025-09-07T09:11:22.1767379Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1768006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1768508Z layer_outputs = layer_module( 2025-09-07T09:11:22.1768870Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1769238Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1769666Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1770090Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1770536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1770957Z self_outputs = self.self( 2025-09-07T09:11:22.1771353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T09:11:22.1771772Z value_vectors = self.value(hidden_states) 2025-09-07T09:11:22.1771911Z 2025-09-07T09:11:22.1772014Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1772533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1773019Z layer_outputs = layer_module( 2025-09-07T09:11:22.1773357Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1773710Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1774210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1774666Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1775070Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1775471Z self_outputs = self.self( 2025-09-07T09:11:22.1775854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1776303Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1776825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1777407Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:11:22.1777935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1778282Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1778430Z 2025-09-07T09:11:22.1778529Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1779038Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1779519Z layer_outputs = layer_module( 2025-09-07T09:11:22.1779889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1780246Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1780654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1781068Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1781472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1781878Z self_outputs = self.self( 2025-09-07T09:11:22.1782261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1782706Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1783227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1783768Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:11:22.1784275Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:11:22.1784732Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:11:22.1785055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1785393Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1785539Z 2025-09-07T09:11:22.1785640Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1786148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1786632Z layer_outputs = layer_module( 2025-09-07T09:11:22.1786972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1787047Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1787335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1787466Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1787752Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1787852Z self_outputs = self.self( 2025-09-07T09:11:22.1788138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1788243Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1788600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1788752Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1788756Z 2025-09-07T09:11:22.1788849Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1789236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1789306Z layer_outputs = layer_module( 2025-09-07T09:11:22.1789530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1789602Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1789881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1789954Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1790262Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1790333Z self_outputs = self.self( 2025-09-07T09:11:22.1790614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1790720Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1791078Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1791215Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1791218Z 2025-09-07T09:11:22.1791312Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1791664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1791734Z layer_outputs = layer_module( 2025-09-07T09:11:22.1791946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1792013Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1792298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1792366Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1792648Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1792712Z self_outputs = self.self( 2025-09-07T09:11:22.1792995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:11:22.1793177Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:11:22.1793180Z 2025-09-07T09:11:22.1793271Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1793626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1793728Z layer_outputs = layer_module( 2025-09-07T09:11:22.1793980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1794051Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1794335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1794406Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1794680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T09:11:22.1794789Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T09:11:22.1795066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T09:11:22.1795148Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1795153Z 2025-09-07T09:11:22.1795278Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1795630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1795699Z layer_outputs = layer_module( 2025-09-07T09:11:22.1795911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1795987Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1796298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1796378Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1796644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1796716Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1797006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1797111Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1797394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T09:11:22.1797467Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1797470Z 2025-09-07T09:11:22.1797565Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1797922Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1797985Z layer_outputs = layer_module( 2025-09-07T09:11:22.1798197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1798267Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1798548Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1798619Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1798880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1798951Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1799232Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1799333Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1799608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T09:11:22.1799716Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T09:11:22.1799981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T09:11:22.1800078Z return self.act(input) 2025-09-07T09:11:22.1800081Z 2025-09-07T09:11:22.1800177Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1800529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1800595Z layer_outputs = layer_module( 2025-09-07T09:11:22.1800808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1800877Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1801162Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1801234Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1801530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1801599Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1801883Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T09:11:22.1802007Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T09:11:22.1802285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T09:11:22.1802451Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1802455Z 2025-09-07T09:11:22.1802546Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1802904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1802970Z layer_outputs = layer_module( 2025-09-07T09:11:22.1803375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1803457Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1803741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1803818Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1804100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1804167Z self_outputs = self.self( 2025-09-07T09:11:22.1804449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:11:22.1804524Z query_vectors = self.query(hidden_states) 2025-09-07T09:11:22.1804530Z 2025-09-07T09:11:22.1804635Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1804991Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1805063Z layer_outputs = layer_module( 2025-09-07T09:11:22.1805284Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1805356Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1805641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1805708Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1805988Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1806139Z self_outputs = self.self( 2025-09-07T09:11:22.1806425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1806574Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1806919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1807113Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1807116Z 2025-09-07T09:11:22.1807212Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1807571Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1807636Z layer_outputs = layer_module( 2025-09-07T09:11:22.1807912Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1807988Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1808269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1808343Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1808619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1808687Z self_outputs = self.self( 2025-09-07T09:11:22.1809032Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T09:11:22.1809103Z key_vectors = self.key(hidden_states) 2025-09-07T09:11:22.1809106Z 2025-09-07T09:11:22.1809200Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1809553Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1809620Z layer_outputs = layer_module( 2025-09-07T09:11:22.1809840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1809910Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1810191Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1810259Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1810541Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1810602Z self_outputs = self.self( 2025-09-07T09:11:22.1810880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1810978Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1811321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1811499Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1811502Z 2025-09-07T09:11:22.1811578Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1811647Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1811719Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1811789Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1811887Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1812236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1812345Z layer_outputs = layer_module( 2025-09-07T09:11:22.1812564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1812669Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1812951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1813018Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1813299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1813364Z self_outputs = self.self( 2025-09-07T09:11:22.1813639Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T09:11:22.1813744Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1814122Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1814272Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T09:11:22.1814596Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T09:11:22.1814742Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T09:11:22.1814745Z 2025-09-07T09:11:22.1814820Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1814946Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1815311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1815380Z layer_outputs = layer_module( 2025-09-07T09:11:22.1815604Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1815677Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1815959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1816035Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1816314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1816381Z self_outputs = self.self( 2025-09-07T09:11:22.1816663Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:11:22.1816742Z attn_scores += diagonal_mask 2025-09-07T09:11:22.1816745Z 2025-09-07T09:11:22.1816837Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1817189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1817263Z layer_outputs = layer_module( 2025-09-07T09:11:22.1817479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1817553Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1817865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1817934Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1818221Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1818285Z self_outputs = self.self( 2025-09-07T09:11:22.1818568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:11:22.1818686Z attn_probs = nn.functional.softmax( 2025-09-07T09:11:22.1818718Z 2025-09-07T09:11:22.1818816Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1819167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1819232Z layer_outputs = layer_module( 2025-09-07T09:11:22.1819453Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1819521Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1819807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1819876Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1820189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1820263Z self_outputs = self.self( 2025-09-07T09:11:22.1820539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T09:11:22.1820624Z value_vectors = self.value(hidden_states) 2025-09-07T09:11:22.1820627Z 2025-09-07T09:11:22.1820720Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1821077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1821176Z layer_outputs = layer_module( 2025-09-07T09:11:22.1821396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1821480Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1821764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1821845Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1822120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1822184Z self_outputs = self.self( 2025-09-07T09:11:22.1822464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1822575Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1822937Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1823105Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:11:22.1823300Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1823393Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1823397Z 2025-09-07T09:11:22.1823492Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1823851Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1823917Z layer_outputs = layer_module( 2025-09-07T09:11:22.1824137Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1824208Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1824492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1824563Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1824879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1824981Z self_outputs = self.self( 2025-09-07T09:11:22.1825258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1825368Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1825722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1825857Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:11:22.1826184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:11:22.1826267Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:11:22.1826495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1826586Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1826589Z 2025-09-07T09:11:22.1826687Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1827045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1827109Z layer_outputs = layer_module( 2025-09-07T09:11:22.1827367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1827438Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1827722Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1827790Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1828078Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1828143Z self_outputs = self.self( 2025-09-07T09:11:22.1828419Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1828526Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1828879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1829031Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1829034Z 2025-09-07T09:11:22.1829126Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1829484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1829550Z layer_outputs = layer_module( 2025-09-07T09:11:22.1829763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1829837Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1830114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1830188Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1830466Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1830527Z self_outputs = self.self( 2025-09-07T09:11:22.1830807Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1830947Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1831303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1831471Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1831475Z 2025-09-07T09:11:22.1831572Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1831924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1831990Z layer_outputs = layer_module( 2025-09-07T09:11:22.1832210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1832280Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1832608Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1832679Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1832958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1833027Z self_outputs = self.self( 2025-09-07T09:11:22.1833302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:11:22.1833485Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:11:22.1833520Z 2025-09-07T09:11:22.1833612Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1833969Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1834033Z layer_outputs = layer_module( 2025-09-07T09:11:22.1834247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1834322Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1834600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1834673Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1834953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T09:11:22.1835065Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T09:11:22.1835349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T09:11:22.1835425Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1835428Z 2025-09-07T09:11:22.1835526Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1835885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1835948Z layer_outputs = layer_module( 2025-09-07T09:11:22.1836160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1836230Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1836509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1836590Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1836852Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1836929Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1837257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1837395Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1837679Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T09:11:22.1837754Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1837757Z 2025-09-07T09:11:22.1837855Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1838206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1838271Z layer_outputs = layer_module( 2025-09-07T09:11:22.1838483Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1838928Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1839213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1839291Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1839554Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1839621Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1839903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1840036Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1840318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T09:11:22.1840427Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T09:11:22.1840640Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T09:11:22.1840707Z return self.act(input) 2025-09-07T09:11:22.1840713Z 2025-09-07T09:11:22.1840806Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1841155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1841222Z layer_outputs = layer_module( 2025-09-07T09:11:22.1841434Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1841505Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1841782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1841854Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1842114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1842180Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1842464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T09:11:22.1842576Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T09:11:22.1842859Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T09:11:22.1842933Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1842936Z 2025-09-07T09:11:22.1843024Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1843371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1843473Z layer_outputs = layer_module( 2025-09-07T09:11:22.1843723Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1843790Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1844064Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1844141Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1844418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1844489Z self_outputs = self.self( 2025-09-07T09:11:22.1844768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:11:22.1844842Z query_vectors = self.query(hidden_states) 2025-09-07T09:11:22.1844846Z 2025-09-07T09:11:22.1844965Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1845321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1845389Z layer_outputs = layer_module( 2025-09-07T09:11:22.1845606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1845679Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1845995Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1846065Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1846349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1846414Z self_outputs = self.self( 2025-09-07T09:11:22.1846697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1846795Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1847143Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1847327Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1847330Z 2025-09-07T09:11:22.1847424Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1847782Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1847848Z layer_outputs = layer_module( 2025-09-07T09:11:22.1848069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1848141Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1848418Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1848487Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1848761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1848826Z self_outputs = self.self( 2025-09-07T09:11:22.1849103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T09:11:22.1849175Z key_vectors = self.key(hidden_states) 2025-09-07T09:11:22.1849178Z 2025-09-07T09:11:22.1849269Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1849620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1849748Z layer_outputs = layer_module( 2025-09-07T09:11:22.1849961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1850037Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1850313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1850381Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1850656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1850718Z self_outputs = self.self( 2025-09-07T09:11:22.1850996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1851119Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1851467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1851642Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1851645Z 2025-09-07T09:11:22.1851725Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1851794Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1851861Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1851964Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1852054Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1852410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1852474Z layer_outputs = layer_module( 2025-09-07T09:11:22.1852694Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1852768Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1853045Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1853116Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1853394Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1853457Z self_outputs = self.self( 2025-09-07T09:11:22.1853737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T09:11:22.1853840Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1854186Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1854328Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T09:11:22.1854654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T09:11:22.1854796Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T09:11:22.1854799Z 2025-09-07T09:11:22.1854871Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1854967Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1855318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1855385Z layer_outputs = layer_module( 2025-09-07T09:11:22.1855598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1855745Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1856029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1856094Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1856372Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1856436Z self_outputs = self.self( 2025-09-07T09:11:22.1856717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:11:22.1856782Z attn_scores += diagonal_mask 2025-09-07T09:11:22.1856785Z 2025-09-07T09:11:22.1856875Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1857261Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1857328Z layer_outputs = layer_module( 2025-09-07T09:11:22.1857545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1857613Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1857947Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1858017Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1858332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1858401Z self_outputs = self.self( 2025-09-07T09:11:22.1858678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:11:22.1858754Z attn_probs = nn.functional.softmax( 2025-09-07T09:11:22.1858759Z 2025-09-07T09:11:22.1858848Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1859195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1859259Z layer_outputs = layer_module( 2025-09-07T09:11:22.1859471Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1859541Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1859817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1859884Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1860160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1860222Z self_outputs = self.self( 2025-09-07T09:11:22.1860501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T09:11:22.1860578Z value_vectors = self.value(hidden_states) 2025-09-07T09:11:22.1860581Z 2025-09-07T09:11:22.1860674Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1861022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1861086Z layer_outputs = layer_module( 2025-09-07T09:11:22.1861301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1861371Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1861652Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1861780Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1862059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1862123Z self_outputs = self.self( 2025-09-07T09:11:22.1862398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1862517Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1862871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1863040Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:11:22.1863260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1863354Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1863359Z 2025-09-07T09:11:22.1863448Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1863797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1863861Z layer_outputs = layer_module( 2025-09-07T09:11:22.1864074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1864190Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1864469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1864537Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1864817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1864880Z self_outputs = self.self( 2025-09-07T09:11:22.1865157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1865261Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1865634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1865765Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:11:22.1866081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:11:22.1866163Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:11:22.1866350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1866437Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1866440Z 2025-09-07T09:11:22.1866529Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1866880Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1866943Z layer_outputs = layer_module( 2025-09-07T09:11:22.1867154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1867225Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1867500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1867567Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1867886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1867979Z self_outputs = self.self( 2025-09-07T09:11:22.1868255Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1868356Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1868707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1868850Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1868853Z 2025-09-07T09:11:22.1868943Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1869327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1869390Z layer_outputs = layer_module( 2025-09-07T09:11:22.1869606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1869674Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1869953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1870018Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1870324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1870390Z self_outputs = self.self( 2025-09-07T09:11:22.1870667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1870770Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1871124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1871264Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1871267Z 2025-09-07T09:11:22.1871358Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1871706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1871773Z layer_outputs = layer_module( 2025-09-07T09:11:22.1871987Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1872057Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1872336Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1872407Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1872682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1872740Z self_outputs = self.self( 2025-09-07T09:11:22.1873019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:11:22.1873194Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:11:22.1873197Z 2025-09-07T09:11:22.1873293Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1873644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1873712Z layer_outputs = layer_module( 2025-09-07T09:11:22.1873964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1874066Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1874344Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1874409Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1874688Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T09:11:22.1874790Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T09:11:22.1875073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T09:11:22.1875148Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1875151Z 2025-09-07T09:11:22.1875241Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1875624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1875688Z layer_outputs = layer_module( 2025-09-07T09:11:22.1875902Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1875969Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1876247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1876359Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1876625Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1876695Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1876981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1877083Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1877360Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T09:11:22.1877433Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1877436Z 2025-09-07T09:11:22.1877528Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1877877Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1877942Z layer_outputs = layer_module( 2025-09-07T09:11:22.1878154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1878220Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1878504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1878576Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1878836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1878901Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1879182Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1879280Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1879556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T09:11:22.1879661Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T09:11:22.1879926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T09:11:22.1880025Z return self.act(input) 2025-09-07T09:11:22.1880028Z 2025-09-07T09:11:22.1880118Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1880472Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1880542Z layer_outputs = layer_module( 2025-09-07T09:11:22.1880753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1880827Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1881103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1881176Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1881468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1881536Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1881823Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T09:11:22.1881941Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T09:11:22.1882222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T09:11:22.1882325Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1882328Z 2025-09-07T09:11:22.1882419Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1882772Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1882837Z layer_outputs = layer_module( 2025-09-07T09:11:22.1883057Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1883124Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1883401Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1883470Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1883746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1883814Z self_outputs = self.self( 2025-09-07T09:11:22.1884088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:11:22.1884165Z query_vectors = self.query(hidden_states) 2025-09-07T09:11:22.1884167Z 2025-09-07T09:11:22.1884258Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1884607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1884669Z layer_outputs = layer_module( 2025-09-07T09:11:22.1884888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1884961Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1885237Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1885304Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1885583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1885646Z self_outputs = self.self( 2025-09-07T09:11:22.1885965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1886090Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1886440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1886629Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1886632Z 2025-09-07T09:11:22.1886729Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1887088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1887153Z layer_outputs = layer_module( 2025-09-07T09:11:22.1887373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1887475Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1887761Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1887831Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1888133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1888200Z self_outputs = self.self( 2025-09-07T09:11:22.1888507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T09:11:22.1888585Z key_vectors = self.key(hidden_states) 2025-09-07T09:11:22.1888588Z 2025-09-07T09:11:22.1888678Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1889036Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1889104Z layer_outputs = layer_module( 2025-09-07T09:11:22.1889313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1889387Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1889662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1889734Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1890010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1890070Z self_outputs = self.self( 2025-09-07T09:11:22.1890349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1890445Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1890792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1890965Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1890968Z 2025-09-07T09:11:22.1891046Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1891115Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1891181Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1891252Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1891344Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1891696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1891758Z layer_outputs = layer_module( 2025-09-07T09:11:22.1892013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1892118Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1892396Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1892465Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1892740Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1892805Z self_outputs = self.self( 2025-09-07T09:11:22.1893083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T09:11:22.1893181Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1893557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1893698Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T09:11:22.1894026Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T09:11:22.1894167Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T09:11:22.1894170Z 2025-09-07T09:11:22.1894243Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1894333Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1894716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1894785Z layer_outputs = layer_module( 2025-09-07T09:11:22.1895001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1895073Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1895349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1895415Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1895697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1895759Z self_outputs = self.self( 2025-09-07T09:11:22.1896040Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:11:22.1896105Z attn_scores += diagonal_mask 2025-09-07T09:11:22.1896108Z 2025-09-07T09:11:22.1896197Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1896552Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1896617Z layer_outputs = layer_module( 2025-09-07T09:11:22.1896832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1896901Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1897184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1897250Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1897526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1897590Z self_outputs = self.self( 2025-09-07T09:11:22.1897945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:11:22.1898055Z attn_probs = nn.functional.softmax( 2025-09-07T09:11:22.1898058Z 2025-09-07T09:11:22.1898176Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1898528Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1898590Z layer_outputs = layer_module( 2025-09-07T09:11:22.1898802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1898872Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1899151Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1899220Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1899495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1899660Z self_outputs = self.self( 2025-09-07T09:11:22.1899946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T09:11:22.1900022Z value_vectors = self.value(hidden_states) 2025-09-07T09:11:22.1900025Z 2025-09-07T09:11:22.1900125Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1900474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1900545Z layer_outputs = layer_module( 2025-09-07T09:11:22.1900804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1900874Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1901161Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1901228Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1901512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1901575Z self_outputs = self.self( 2025-09-07T09:11:22.1901849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1901969Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1902326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1902501Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:11:22.1902689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1902788Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1902793Z 2025-09-07T09:11:22.1902885Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1903364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1903436Z layer_outputs = layer_module( 2025-09-07T09:11:22.1903654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1903731Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1904014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1904089Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1904373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1904552Z self_outputs = self.self( 2025-09-07T09:11:22.1904840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1904950Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1905308Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1905437Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:11:22.1905758Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:11:22.1905842Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:11:22.1906080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1906177Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1906181Z 2025-09-07T09:11:22.1906278Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1906636Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1906703Z layer_outputs = layer_module( 2025-09-07T09:11:22.1906918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1907037Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1907319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1907394Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1907673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1907740Z self_outputs = self.self( 2025-09-07T09:11:22.1908023Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1908127Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1908485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1908635Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1908639Z 2025-09-07T09:11:22.1908737Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1909091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1909161Z layer_outputs = layer_module( 2025-09-07T09:11:22.1909382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1909449Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1909734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1909803Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1910089Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1910149Z self_outputs = self.self( 2025-09-07T09:11:22.1910426Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1910535Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1910934Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1911113Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1911116Z 2025-09-07T09:11:22.1911210Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1911569Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1911635Z layer_outputs = layer_module( 2025-09-07T09:11:22.1911849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1911928Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1912205Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1912314Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1912603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1912668Z self_outputs = self.self( 2025-09-07T09:11:22.1912955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:11:22.1913135Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:11:22.1913138Z 2025-09-07T09:11:22.1913266Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1913620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1913691Z layer_outputs = layer_module( 2025-09-07T09:11:22.1913909Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1913981Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1914264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1914333Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1914614Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T09:11:22.1914720Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T09:11:22.1914997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T09:11:22.1915078Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1915081Z 2025-09-07T09:11:22.1915171Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1915529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1915594Z layer_outputs = layer_module( 2025-09-07T09:11:22.1915813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1915879Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1916158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1916241Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1916503Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1916579Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1916862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1917033Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1917315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T09:11:22.1917389Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1917392Z 2025-09-07T09:11:22.1917489Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1917840Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1917907Z layer_outputs = layer_module( 2025-09-07T09:11:22.1918118Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1918188Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1918507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1918584Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1918848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1918914Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1919202Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1919331Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1919612Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T09:11:22.1919722Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T09:11:22.1919932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T09:11:22.1920004Z return self.act(input) 2025-09-07T09:11:22.1920007Z 2025-09-07T09:11:22.1920098Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1920456Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1920519Z layer_outputs = layer_module( 2025-09-07T09:11:22.1920732Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1920809Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1921087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1921168Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1921427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1921495Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1921781Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T09:11:22.1921893Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T09:11:22.1922178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T09:11:22.1922251Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1922256Z 2025-09-07T09:11:22.1922352Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1922701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1922799Z layer_outputs = layer_module( 2025-09-07T09:11:22.1923017Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1923124Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1923404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1923471Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1923747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1923817Z self_outputs = self.self( 2025-09-07T09:11:22.1924095Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:11:22.1924174Z query_vectors = self.query(hidden_states) 2025-09-07T09:11:22.1924177Z 2025-09-07T09:11:22.1924267Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1924672Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1924739Z layer_outputs = layer_module( 2025-09-07T09:11:22.1924955Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1925028Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1925305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1925409Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1925689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1925754Z self_outputs = self.self( 2025-09-07T09:11:22.1926038Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1926133Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1926485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1926667Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1926670Z 2025-09-07T09:11:22.1926766Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1927117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1927186Z layer_outputs = layer_module( 2025-09-07T09:11:22.1927410Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1927483Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1927768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1927835Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1928117Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1928180Z self_outputs = self.self( 2025-09-07T09:11:22.1928459Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T09:11:22.1928539Z key_vectors = self.key(hidden_states) 2025-09-07T09:11:22.1928542Z 2025-09-07T09:11:22.1928629Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1928984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1929112Z layer_outputs = layer_module( 2025-09-07T09:11:22.1929326Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1929400Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1929678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1929749Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1930029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1930095Z self_outputs = self.self( 2025-09-07T09:11:22.1930375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1930464Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1930861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1931049Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1931053Z 2025-09-07T09:11:22.1931136Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1931205Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1931274Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1931350Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1931478Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1931854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1931923Z layer_outputs = layer_module( 2025-09-07T09:11:22.1932155Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1932229Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1932514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1932594Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1932873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1932940Z self_outputs = self.self( 2025-09-07T09:11:22.1933219Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T09:11:22.1933321Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1933671Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1933814Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T09:11:22.1934150Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T09:11:22.1934294Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T09:11:22.1934297Z 2025-09-07T09:11:22.1934375Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1934468Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1934825Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1934896Z layer_outputs = layer_module( 2025-09-07T09:11:22.1935112Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1935225Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1935538Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1935609Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1935894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1935959Z self_outputs = self.self( 2025-09-07T09:11:22.1936240Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:11:22.1936304Z attn_scores += diagonal_mask 2025-09-07T09:11:22.1936308Z 2025-09-07T09:11:22.1936405Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1936790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1936858Z layer_outputs = layer_module( 2025-09-07T09:11:22.1937079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1937147Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1937430Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1937498Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1937864Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1937928Z self_outputs = self.self( 2025-09-07T09:11:22.1938216Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:11:22.1938295Z attn_probs = nn.functional.softmax( 2025-09-07T09:11:22.1938302Z 2025-09-07T09:11:22.1938392Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1938751Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1938814Z layer_outputs = layer_module( 2025-09-07T09:11:22.1939028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1939102Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1939382Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1939453Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1939728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1939800Z self_outputs = self.self( 2025-09-07T09:11:22.1940076Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T09:11:22.1940153Z value_vectors = self.value(hidden_states) 2025-09-07T09:11:22.1940157Z 2025-09-07T09:11:22.1940254Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1940603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1940671Z layer_outputs = layer_module( 2025-09-07T09:11:22.1940886Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1940954Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1941234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1941339Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1941656Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1941719Z self_outputs = self.self( 2025-09-07T09:11:22.1942002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1942114Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1942473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1942647Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:11:22.1942836Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1942964Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1942970Z 2025-09-07T09:11:22.1943062Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1943424Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1943488Z layer_outputs = layer_module( 2025-09-07T09:11:22.1943702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1943774Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1944086Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1944158Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1944439Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1944503Z self_outputs = self.self( 2025-09-07T09:11:22.1944786Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1944891Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1945250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1945377Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:11:22.1945703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:11:22.1945786Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:11:22.1945967Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1946061Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1946066Z 2025-09-07T09:11:22.1946157Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1946514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1946578Z layer_outputs = layer_module( 2025-09-07T09:11:22.1946792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1946867Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1947144Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1947216Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1947493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1947634Z self_outputs = self.self( 2025-09-07T09:11:22.1947908Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1948008Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1948363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1948507Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1948511Z 2025-09-07T09:11:22.1948607Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1948959Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1949058Z layer_outputs = layer_module( 2025-09-07T09:11:22.1949274Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1949342Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1949626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1949694Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1949974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1950066Z self_outputs = self.self( 2025-09-07T09:11:22.1950342Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1950449Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1950803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1950941Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1950944Z 2025-09-07T09:11:22.1951035Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1951388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1951454Z layer_outputs = layer_module( 2025-09-07T09:11:22.1951669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1951745Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1952023Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1952098Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1952375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1952447Z self_outputs = self.self( 2025-09-07T09:11:22.1952720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:11:22.1952900Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:11:22.1952903Z 2025-09-07T09:11:22.1953000Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1953353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1953419Z layer_outputs = layer_module( 2025-09-07T09:11:22.1953632Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1953764Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1954050Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1954118Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1954402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T09:11:22.1954505Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T09:11:22.1954790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T09:11:22.1954865Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1954868Z 2025-09-07T09:11:22.1954957Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1955346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1955413Z layer_outputs = layer_module( 2025-09-07T09:11:22.1955634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1955702Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1955984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1956092Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1956361Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1956437Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1956724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1956834Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1957108Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T09:11:22.1957184Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1957191Z 2025-09-07T09:11:22.1957282Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1957643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1957711Z layer_outputs = layer_module( 2025-09-07T09:11:22.1957925Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1958002Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1958285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1958358Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1958621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1958689Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1958974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1959070Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1959349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T09:11:22.1959458Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T09:11:22.1959669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T09:11:22.1959779Z return self.act(input) 2025-09-07T09:11:22.1959812Z 2025-09-07T09:11:22.1959907Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1960264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1960327Z layer_outputs = layer_module( 2025-09-07T09:11:22.1960538Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1960616Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1960895Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1960974Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1961260Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1961332Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1961621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T09:11:22.1961735Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T09:11:22.1962020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T09:11:22.1962092Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1962096Z 2025-09-07T09:11:22.1962223Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1962577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1962642Z layer_outputs = layer_module( 2025-09-07T09:11:22.1962865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1962935Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1963218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1963287Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1963575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1963639Z self_outputs = self.self( 2025-09-07T09:11:22.1963915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:11:22.1963995Z query_vectors = self.query(hidden_states) 2025-09-07T09:11:22.1963998Z 2025-09-07T09:11:22.1964088Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1964450Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1964520Z layer_outputs = layer_module( 2025-09-07T09:11:22.1964741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1964815Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1965092Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1965170Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1965446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1965515Z self_outputs = self.self( 2025-09-07T09:11:22.1965791Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1965952Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1966306Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1966488Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1966491Z 2025-09-07T09:11:22.1966591Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1966943Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1967014Z layer_outputs = layer_module( 2025-09-07T09:11:22.1967227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1967332Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1967624Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1967695Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1967979Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1968043Z self_outputs = self.self( 2025-09-07T09:11:22.1968321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T09:11:22.1968432Z key_vectors = self.key(hidden_states) 2025-09-07T09:11:22.1968435Z 2025-09-07T09:11:22.1968526Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1968882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1968951Z layer_outputs = layer_module( 2025-09-07T09:11:22.1969181Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1969252Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1969530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1969607Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1969887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1969955Z self_outputs = self.self( 2025-09-07T09:11:22.1970230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.1970321Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1970678Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1970852Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.1970855Z 2025-09-07T09:11:22.1970933Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1971005Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1971080Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1971150Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1971241Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1971601Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1971670Z layer_outputs = layer_module( 2025-09-07T09:11:22.1971894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1972041Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1972319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1972394Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1972669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1972737Z self_outputs = self.self( 2025-09-07T09:11:22.1973015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T09:11:22.1973122Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.1973465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.1973641Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T09:11:22.1973978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T09:11:22.1974118Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T09:11:22.1974122Z 2025-09-07T09:11:22.1974199Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.1974291Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1974689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1974754Z layer_outputs = layer_module( 2025-09-07T09:11:22.1974968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1975048Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1975329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1975405Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1975681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1975746Z self_outputs = self.self( 2025-09-07T09:11:22.1976030Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:11:22.1976098Z attn_scores += diagonal_mask 2025-09-07T09:11:22.1976102Z 2025-09-07T09:11:22.1976199Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1976551Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1976623Z layer_outputs = layer_module( 2025-09-07T09:11:22.1976842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1976911Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1977194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1977261Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1977546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1977608Z self_outputs = self.self( 2025-09-07T09:11:22.1977928Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:11:22.1978007Z attn_probs = nn.functional.softmax( 2025-09-07T09:11:22.1978045Z 2025-09-07T09:11:22.1978139Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1978536Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1978602Z layer_outputs = layer_module( 2025-09-07T09:11:22.1978828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1978900Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1979178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1979254Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1979534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1979607Z self_outputs = self.self( 2025-09-07T09:11:22.1979917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T09:11:22.1979996Z value_vectors = self.value(hidden_states) 2025-09-07T09:11:22.1979999Z 2025-09-07T09:11:22.1980096Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1980449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1980521Z layer_outputs = layer_module( 2025-09-07T09:11:22.1980779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1980856Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1981142Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1981217Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1981504Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1981570Z self_outputs = self.self( 2025-09-07T09:11:22.1981850Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1981961Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1982330Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1982503Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:11:22.1982691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1982790Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1982793Z 2025-09-07T09:11:22.1982887Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1983246Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1983309Z layer_outputs = layer_module( 2025-09-07T09:11:22.1983521Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1983598Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1983873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1983946Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1984220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1984331Z self_outputs = self.self( 2025-09-07T09:11:22.1984645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1984749Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1985110Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1985240Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:11:22.1985568Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:11:22.1985651Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:11:22.1985833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.1985959Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.1985964Z 2025-09-07T09:11:22.1986055Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1986411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1986473Z layer_outputs = layer_module( 2025-09-07T09:11:22.1986692Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1986760Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1987073Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1987149Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1987431Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1987504Z self_outputs = self.self( 2025-09-07T09:11:22.1987784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1987886Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1988243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1988388Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1988392Z 2025-09-07T09:11:22.1988489Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1988842Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1988912Z layer_outputs = layer_module( 2025-09-07T09:11:22.1989126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1989197Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1989481Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1989549Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1989832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1989896Z self_outputs = self.self( 2025-09-07T09:11:22.1990179Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.1990279Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.1990627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.1990837Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.1990840Z 2025-09-07T09:11:22.1990935Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1991293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1991356Z layer_outputs = layer_module( 2025-09-07T09:11:22.1991576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1991644Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1991928Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1992035Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1992319Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.1992390Z self_outputs = self.self( 2025-09-07T09:11:22.1992667Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:11:22.1992844Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:11:22.1992854Z 2025-09-07T09:11:22.1992947Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1993348Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1993419Z layer_outputs = layer_module( 2025-09-07T09:11:22.1993635Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1993713Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1993989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.1994057Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.1994338Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T09:11:22.1994440Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T09:11:22.1994728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T09:11:22.1994805Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1994808Z 2025-09-07T09:11:22.1994902Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1995250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1995317Z layer_outputs = layer_module( 2025-09-07T09:11:22.1995533Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1995599Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1995882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1995956Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1996215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1996289Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1996572Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1996717Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1997081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T09:11:22.1997165Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.1997168Z 2025-09-07T09:11:22.1997260Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.1997610Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.1997682Z layer_outputs = layer_module( 2025-09-07T09:11:22.1997894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.1997970Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.1998277Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.1998354Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.1998618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.1998685Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.1998973Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.1999069Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.1999387Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T09:11:22.1999490Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T09:11:22.1999700Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T09:11:22.1999770Z return self.act(input) 2025-09-07T09:11:22.1999775Z 2025-09-07T09:11:22.1999867Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2000226Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2000290Z layer_outputs = layer_module( 2025-09-07T09:11:22.2000509Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2000578Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2000854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.2000935Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.2001188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.2001265Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.2001550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T09:11:22.2001663Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T09:11:22.2001945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T09:11:22.2002019Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.2002022Z 2025-09-07T09:11:22.2002118Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2002469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2002537Z layer_outputs = layer_module( 2025-09-07T09:11:22.2002790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2002893Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2003305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2003376Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2003662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2003724Z self_outputs = self.self( 2025-09-07T09:11:22.2004006Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:11:22.2004088Z query_vectors = self.query(hidden_states) 2025-09-07T09:11:22.2004091Z 2025-09-07T09:11:22.2004181Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2004590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2004657Z layer_outputs = layer_module( 2025-09-07T09:11:22.2004875Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2004943Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2005220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2005295Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2005622Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2005692Z self_outputs = self.self( 2025-09-07T09:11:22.2005970Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.2006068Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.2006421Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.2006629Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.2006632Z 2025-09-07T09:11:22.2006733Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2007083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2007154Z layer_outputs = layer_module( 2025-09-07T09:11:22.2007366Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2007435Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2007717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2007784Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2008067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2008130Z self_outputs = self.self( 2025-09-07T09:11:22.2008411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T09:11:22.2008482Z key_vectors = self.key(hidden_states) 2025-09-07T09:11:22.2008485Z 2025-09-07T09:11:22.2008575Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2008929Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2009040Z layer_outputs = layer_module( 2025-09-07T09:11:22.2009314Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2009384Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2009668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2009743Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2010019Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2010088Z self_outputs = self.self( 2025-09-07T09:11:22.2010371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.2010473Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.2010848Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.2011033Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.2011036Z 2025-09-07T09:11:22.2011115Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.2011185Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.2011259Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.2011327Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.2011425Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2011817Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2011883Z layer_outputs = layer_module( 2025-09-07T09:11:22.2012106Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2012176Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2012463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2012528Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2012803Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2012871Z self_outputs = self.self( 2025-09-07T09:11:22.2013149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T09:11:22.2013252Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.2013598Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.2013736Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T09:11:22.2014065Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T09:11:22.2014206Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T09:11:22.2014209Z 2025-09-07T09:11:22.2014282Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.2014375Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2014728Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2014790Z layer_outputs = layer_module( 2025-09-07T09:11:22.2015002Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2015111Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2015390Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2015497Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2015778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2015842Z self_outputs = self.self( 2025-09-07T09:11:22.2016120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:11:22.2016184Z attn_scores += diagonal_mask 2025-09-07T09:11:22.2016187Z 2025-09-07T09:11:22.2016282Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2016630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2016748Z layer_outputs = layer_module( 2025-09-07T09:11:22.2016965Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2017036Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2017318Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2017385Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2017699Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2017829Z self_outputs = self.self( 2025-09-07T09:11:22.2018113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:11:22.2018184Z attn_probs = nn.functional.softmax( 2025-09-07T09:11:22.2018190Z 2025-09-07T09:11:22.2018285Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2018643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2018709Z layer_outputs = layer_module( 2025-09-07T09:11:22.2018924Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2018991Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2019270Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2019341Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2019617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2019684Z self_outputs = self.self( 2025-09-07T09:11:22.2019963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T09:11:22.2020044Z value_vectors = self.value(hidden_states) 2025-09-07T09:11:22.2020046Z 2025-09-07T09:11:22.2020136Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2020484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2020552Z layer_outputs = layer_module( 2025-09-07T09:11:22.2020767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2020841Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2021116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2021220Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2021506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2021601Z self_outputs = self.self( 2025-09-07T09:11:22.2021882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.2021992Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.2022354Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.2022519Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:11:22.2022709Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.2022805Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.2022841Z 2025-09-07T09:11:22.2022936Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2023298Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2023361Z layer_outputs = layer_module( 2025-09-07T09:11:22.2023578Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2023647Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2023957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2024031Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2024313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2024384Z self_outputs = self.self( 2025-09-07T09:11:22.2024659Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.2024763Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.2025124Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.2025250Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:11:22.2025574Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:11:22.2025656Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:11:22.2025846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.2025933Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.2025937Z 2025-09-07T09:11:22.2026028Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2026385Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2026449Z layer_outputs = layer_module( 2025-09-07T09:11:22.2026668Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2026733Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2027010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2027094Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2027370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2027475Z self_outputs = self.self( 2025-09-07T09:11:22.2027789Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.2027898Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.2028248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.2028391Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.2028395Z 2025-09-07T09:11:22.2028494Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2028843Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2028913Z layer_outputs = layer_module( 2025-09-07T09:11:22.2029158Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2029237Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2033474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2033577Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2033904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2033972Z self_outputs = self.self( 2025-09-07T09:11:22.2034345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.2034464Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.2034839Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.2034995Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.2035000Z 2025-09-07T09:11:22.2035100Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2035467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2035534Z layer_outputs = layer_module( 2025-09-07T09:11:22.2035767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2035849Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2036138Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2036213Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2036495Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2036567Z self_outputs = self.self( 2025-09-07T09:11:22.2036849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:11:22.2037032Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:11:22.2037041Z 2025-09-07T09:11:22.2037140Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2037501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2037571Z layer_outputs = layer_module( 2025-09-07T09:11:22.2037792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2037908Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2038223Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2038294Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2038576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T09:11:22.2038681Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T09:11:22.2038961Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T09:11:22.2039038Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.2039042Z 2025-09-07T09:11:22.2039142Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2039526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2039594Z layer_outputs = layer_module( 2025-09-07T09:11:22.2039815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2039886Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2040167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.2040241Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.2040542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.2040615Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.2040903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.2041015Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.2041292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T09:11:22.2041370Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.2041374Z 2025-09-07T09:11:22.2041467Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2041818Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2041888Z layer_outputs = layer_module( 2025-09-07T09:11:22.2042100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2042171Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2042447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.2042522Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.2042790Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.2042856Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.2043140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.2043234Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.2043517Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T09:11:22.2043622Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T09:11:22.2043832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T09:11:22.2043940Z return self.act(input) 2025-09-07T09:11:22.2043943Z 2025-09-07T09:11:22.2044085Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2044440Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2044504Z layer_outputs = layer_module( 2025-09-07T09:11:22.2044715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2044788Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2045067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.2045142Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.2045400Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.2045505Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.2045792Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T09:11:22.2045907Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T09:11:22.2046190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T09:11:22.2046265Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.2046269Z 2025-09-07T09:11:22.2046397Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2046756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2046828Z layer_outputs = layer_module( 2025-09-07T09:11:22.2047044Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2047119Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2047399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2047468Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2047753Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2047819Z self_outputs = self.self( 2025-09-07T09:11:22.2048100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 509, in forward 2025-09-07T09:11:22.2048182Z query_vectors = self.query(hidden_states) 2025-09-07T09:11:22.2048186Z 2025-09-07T09:11:22.2048276Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2048630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2048698Z layer_outputs = layer_module( 2025-09-07T09:11:22.2048913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2048980Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2049254Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2049326Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2049603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2049668Z self_outputs = self.self( 2025-09-07T09:11:22.2049948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.2050080Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.2050461Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.2050648Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.2050652Z 2025-09-07T09:11:22.2050749Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2051100Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2051169Z layer_outputs = layer_module( 2025-09-07T09:11:22.2051380Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2051448Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2051765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2051836Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2052119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2052182Z self_outputs = self.self( 2025-09-07T09:11:22.2052460Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 510, in forward 2025-09-07T09:11:22.2052531Z key_vectors = self.key(hidden_states) 2025-09-07T09:11:22.2052566Z 2025-09-07T09:11:22.2052658Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2053016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2053082Z layer_outputs = layer_module( 2025-09-07T09:11:22.2053301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2053373Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2053655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2053725Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2054001Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2054068Z self_outputs = self.self( 2025-09-07T09:11:22.2054345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 524, in forward 2025-09-07T09:11:22.2054439Z attn_scores = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.2054784Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 796, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.2054959Z diagonal_chunked_attention_scores = torch.einsum("bcxd,bcyd->bcxy", (query, key)) # multiply 2025-09-07T09:11:22.2054963Z 2025-09-07T09:11:22.2055044Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.2055113Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.2055182Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.2055249Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.2055339Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2055693Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2055756Z layer_outputs = layer_module( 2025-09-07T09:11:22.2055974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2056082Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2056408Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2056475Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2056759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2056825Z self_outputs = self.self( 2025-09-07T09:11:22.2057101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 536, in forward 2025-09-07T09:11:22.2057202Z diagonal_mask = self._sliding_chunks_query_key_matmul( 2025-09-07T09:11:22.2057545Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 834, in _sliding_chunks_query_key_matmul 2025-09-07T09:11:22.2057719Z self._mask_invalid_locations(diagonal_attention_scores, window_overlap) 2025-09-07T09:11:22.2058119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 762, in _mask_invalid_locations 2025-09-07T09:11:22.2058264Z input_tensor[:, :affected_seq_len, :, : affected_seq_len + 1] = torch.full_like( 2025-09-07T09:11:22.2058269Z 2025-09-07T09:11:22.2058356Z cudagraph partition due to non gpu ops 2025-09-07T09:11:22.2058461Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2058863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2058927Z layer_outputs = layer_module( 2025-09-07T09:11:22.2059148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2059224Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2059507Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2059583Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2059862Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2059927Z self_outputs = self.self( 2025-09-07T09:11:22.2060206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 541, in forward 2025-09-07T09:11:22.2060273Z attn_scores += diagonal_mask 2025-09-07T09:11:22.2060277Z 2025-09-07T09:11:22.2060371Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2060719Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2060785Z layer_outputs = layer_module( 2025-09-07T09:11:22.2060997Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2061066Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2061346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2061413Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2061691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2061753Z self_outputs = self.self( 2025-09-07T09:11:22.2062025Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 579, in forward 2025-09-07T09:11:22.2062096Z attn_probs = nn.functional.softmax( 2025-09-07T09:11:22.2062100Z 2025-09-07T09:11:22.2062226Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2062579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2062672Z layer_outputs = layer_module( 2025-09-07T09:11:22.2062887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2062955Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2063230Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2063299Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2063576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2063638Z self_outputs = self.self( 2025-09-07T09:11:22.2063949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 511, in forward 2025-09-07T09:11:22.2064031Z value_vectors = self.value(hidden_states) 2025-09-07T09:11:22.2064034Z 2025-09-07T09:11:22.2064123Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2064470Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2064535Z layer_outputs = layer_module( 2025-09-07T09:11:22.2064778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2064850Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2065126Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2065192Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2065474Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2065538Z self_outputs = self.self( 2025-09-07T09:11:22.2065813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.2065923Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.2066282Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 863, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.2066453Z padded_value = nn.functional.pad(value, (0, 0, window_overlap, window_overlap), value=-1) 2025-09-07T09:11:22.2066645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.2066738Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.2066742Z 2025-09-07T09:11:22.2066835Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2067192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2067258Z layer_outputs = layer_module( 2025-09-07T09:11:22.2067479Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2067554Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2067835Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2067911Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2068188Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2068306Z self_outputs = self.self( 2025-09-07T09:11:22.2068585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.2068727Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.2069089Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 876, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.2069217Z chunked_attn_probs = self._pad_and_diagonalize(chunked_attn_probs) 2025-09-07T09:11:22.2069547Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 699, in _pad_and_diagonalize 2025-09-07T09:11:22.2069631Z chunked_hidden_states = nn.functional.pad( 2025-09-07T09:11:22.2069822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/torch/nn/functional.py", line 5294, in pad 2025-09-07T09:11:22.2069908Z return torch._C._nn.pad(input, pad, mode, value) 2025-09-07T09:11:22.2069945Z 2025-09-07T09:11:22.2070038Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2070399Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2070462Z layer_outputs = layer_module( 2025-09-07T09:11:22.2070684Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2070752Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2071063Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2071138Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2071416Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2071490Z self_outputs = self.self( 2025-09-07T09:11:22.2071764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.2071872Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.2072236Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.2072381Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.2072385Z 2025-09-07T09:11:22.2072484Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2072833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2072897Z layer_outputs = layer_module( 2025-09-07T09:11:22.2073119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2073187Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2073469Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2073538Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2073813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2073883Z self_outputs = self.self( 2025-09-07T09:11:22.2074160Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 613, in forward 2025-09-07T09:11:22.2074268Z attn_output = self._sliding_chunks_matmul_attn_probs_value( 2025-09-07T09:11:22.2074621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 878, in _sliding_chunks_matmul_attn_probs_value 2025-09-07T09:11:22.2074829Z context = torch.einsum("bcwd,bcdh->bcwh", (chunked_attn_probs, chunked_value)) 2025-09-07T09:11:22.2074833Z 2025-09-07T09:11:22.2074922Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2075278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2075345Z layer_outputs = layer_module( 2025-09-07T09:11:22.2075559Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2075633Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2075915Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2075987Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2076310Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1135, in forward 2025-09-07T09:11:22.2076375Z self_outputs = self.self( 2025-09-07T09:11:22.2076655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 618, in forward 2025-09-07T09:11:22.2076835Z attn_output = attn_output.transpose(0, 1).reshape(seq_len, batch_size, embed_dim).contiguous() 2025-09-07T09:11:22.2076838Z 2025-09-07T09:11:22.2076934Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2077512Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2077582Z layer_outputs = layer_module( 2025-09-07T09:11:22.2077797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2077868Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2078145Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1199, in forward 2025-09-07T09:11:22.2078211Z self_attn_outputs = self.attention( 2025-09-07T09:11:22.2078492Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1144, in forward 2025-09-07T09:11:22.2078595Z attn_output = self.output(self_outputs[0], hidden_states) 2025-09-07T09:11:22.2078871Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1094, in forward 2025-09-07T09:11:22.2078948Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.2078952Z 2025-09-07T09:11:22.2079040Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2079392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2079456Z layer_outputs = layer_module( 2025-09-07T09:11:22.2079670Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2079738Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2080015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.2080090Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.2080352Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.2080426Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.2080707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.2080847Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.2081165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1160, in forward 2025-09-07T09:11:22.2081238Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.2081241Z 2025-09-07T09:11:22.2081335Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2081685Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2081749Z layer_outputs = layer_module( 2025-09-07T09:11:22.2081968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2082036Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2082321Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.2082425Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.2082689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.2082757Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.2083041Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1218, in ff_chunk 2025-09-07T09:11:22.2083137Z intermediate_output = self.intermediate(attn_output) 2025-09-07T09:11:22.2083447Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1161, in forward 2025-09-07T09:11:22.2083556Z hidden_states = self.intermediate_act_fn(hidden_states) 2025-09-07T09:11:22.2083765Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/activations.py", line 69, in forward 2025-09-07T09:11:22.2083834Z return self.act(input) 2025-09-07T09:11:22.2083839Z 2025-09-07T09:11:22.2083928Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:11:22.2084278Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1259, in torch_dynamo_resume_in_forward_at_1244 2025-09-07T09:11:22.2084343Z layer_outputs = layer_module( 2025-09-07T09:11:22.2084558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/modeling_layers.py", line 94, in __call__ 2025-09-07T09:11:22.2084627Z return super().__call__(*args, **kwargs) 2025-09-07T09:11:22.2084904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1211, in forward 2025-09-07T09:11:22.2084978Z layer_output = apply_chunking_to_forward( 2025-09-07T09:11:22.2085233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:11:22.2085301Z return forward_fn(*input_tensors) 2025-09-07T09:11:22.2085583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1219, in ff_chunk 2025-09-07T09:11:22.2085697Z layer_output = self.output(intermediate_output, attn_output) 2025-09-07T09:11:22.2085977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1174, in forward 2025-09-07T09:11:22.2086047Z hidden_states = self.dense(hidden_states) 2025-09-07T09:11:22.2086051Z 2025-09-07T09:12:01.6002155Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:01.6002828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1716, in torch_dynamo_resume_in_forward_at_1703 2025-09-07T09:12:01.6003513Z prediction_scores = self.lm_head(sequence_output) 2025-09-07T09:12:01.6003984Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1333, in forward 2025-09-07T09:12:01.6005567Z x = self.dense(features) 2025-09-07T09:12:01.6005704Z 2025-09-07T09:12:01.6005809Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:01.6006345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1716, in torch_dynamo_resume_in_forward_at_1703 2025-09-07T09:12:01.6006884Z prediction_scores = self.lm_head(sequence_output) 2025-09-07T09:12:01.6007340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/longformer/modeling_longformer.py", line 1338, in forward 2025-09-07T09:12:01.6007780Z x = self.decoder(x) 2025-09-07T09:12:01.6007892Z 2025-09-07T09:12:02.8881625Z 2025-09-07T09:12:03.1225587Z running benchmark: 0% 0/30 [00:00 2025-09-07T09:12:25.9049241Z torch.index_select(position_encodings[i], 0, position_ids[i]).unsqueeze(0) 2025-09-07T09:12:25.9049460Z 2025-09-07T09:12:26.8214366Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:26.8215005Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 370, in torch_dynamo_resume_in_forward_at_369 2025-09-07T09:12:26.8215521Z embeddings = embeddings + position_embeddings 2025-09-07T09:12:26.8215680Z 2025-09-07T09:12:30.4281158Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4281488Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4282101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4282649Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4283114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4283546Z layer_outputs = layer( 2025-09-07T09:12:30.4283932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4284346Z attn_outputs = self.attention( 2025-09-07T09:12:30.4284756Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4285185Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4285616Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:12:30.4286484Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:12:30.4287199Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 807, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:12:30.4287752Z undo_sorted_bucket_idx.scatter_(-1, sorted_bucket_idx, indices) 2025-09-07T09:12:30.4287932Z 2025-09-07T09:12:30.4288043Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4288563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4289081Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4289506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4289994Z layer_outputs = layer( 2025-09-07T09:12:30.4290381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4290783Z attn_outputs = self.attention( 2025-09-07T09:12:30.4291178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4291598Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4292016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:12:30.4292608Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:12:30.4293220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 807, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:12:30.4293774Z undo_sorted_bucket_idx.scatter_(-1, sorted_bucket_idx, indices) 2025-09-07T09:12:30.4293958Z 2025-09-07T09:12:30.4294061Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4294566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4295147Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4295558Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4295955Z layer_outputs = layer( 2025-09-07T09:12:30.4296335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4296748Z attn_outputs = self.attention( 2025-09-07T09:12:30.4297140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4297557Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4298164Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:12:30.4298689Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:12:30.4299304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 807, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:12:30.4299851Z undo_sorted_bucket_idx.scatter_(-1, sorted_bucket_idx, indices) 2025-09-07T09:12:30.4300024Z 2025-09-07T09:12:30.4300124Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4300627Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4301175Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4301705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4302115Z layer_outputs = layer( 2025-09-07T09:12:30.4302486Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4302891Z attn_outputs = self.attention( 2025-09-07T09:12:30.4303404Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4303834Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4304252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1198, in forward 2025-09-07T09:12:30.4304656Z query_vectors = self.query(hidden_states) 2025-09-07T09:12:30.4304807Z 2025-09-07T09:12:30.4304960Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4305464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4305956Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4306364Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4306757Z layer_outputs = layer( 2025-09-07T09:12:30.4307171Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4307581Z attn_outputs = self.attention( 2025-09-07T09:12:30.4307975Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4308399Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4308805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1199, in forward 2025-09-07T09:12:30.4309209Z key_vectors = self.key(hidden_states) 2025-09-07T09:12:30.4309347Z 2025-09-07T09:12:30.4309424Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4309627Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4309840Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4310337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4310826Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4311242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4311642Z layer_outputs = layer( 2025-09-07T09:12:30.4312014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4312418Z attn_outputs = self.attention( 2025-09-07T09:12:30.4312811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4313229Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4313654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T09:12:30.4314185Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4314731Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:12:30.4315233Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:12:30.4315497Z 2025-09-07T09:12:30.4315587Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4315836Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4316050Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4316561Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4317057Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4317478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4317880Z layer_outputs = layer( 2025-09-07T09:12:30.4318245Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4318652Z attn_outputs = self.attention( 2025-09-07T09:12:30.4319080Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4319508Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4319918Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T09:12:30.4320423Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4320956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:12:30.4321393Z return torch.cat(slices, dim=3) 2025-09-07T09:12:30.4321558Z 2025-09-07T09:12:30.4321664Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4322177Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4322672Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4323098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4323513Z layer_outputs = layer( 2025-09-07T09:12:30.4323896Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4324300Z attn_outputs = self.attention( 2025-09-07T09:12:30.4324689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4325118Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4325534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T09:12:30.4326026Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:12:30.4326237Z 2025-09-07T09:12:30.4326342Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4326860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4327357Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4327771Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4328170Z layer_outputs = layer( 2025-09-07T09:12:30.4328555Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4328959Z attn_outputs = self.attention( 2025-09-07T09:12:30.4329350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4329776Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4330242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T09:12:30.4330759Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:12:30.4330971Z 2025-09-07T09:12:30.4331071Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4331580Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4332073Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4332493Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4332881Z layer_outputs = layer( 2025-09-07T09:12:30.4333252Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4333725Z attn_outputs = self.attention( 2025-09-07T09:12:30.4334116Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4334530Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4334935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1200, in forward 2025-09-07T09:12:30.4335349Z value_vectors = self.value(hidden_states) 2025-09-07T09:12:30.4335492Z 2025-09-07T09:12:30.4335569Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4335821Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4336036Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4336537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4337023Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4337435Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4337904Z layer_outputs = layer( 2025-09-07T09:12:30.4338269Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4338671Z attn_outputs = self.attention( 2025-09-07T09:12:30.4339059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4339477Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4339885Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1266, in forward 2025-09-07T09:12:30.4340401Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4340949Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:12:30.4341450Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:12:30.4341657Z 2025-09-07T09:12:30.4341737Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4341940Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4342152Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4342651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4343140Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4343550Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4343998Z layer_outputs = layer( 2025-09-07T09:12:30.4344367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4344807Z attn_outputs = self.attention( 2025-09-07T09:12:30.4345194Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4345610Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4346015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T09:12:30.4346460Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:12:30.4346635Z 2025-09-07T09:12:30.4346731Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4347227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4347748Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4348159Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4348564Z layer_outputs = layer( 2025-09-07T09:12:30.4348938Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4349341Z attn_outputs = self.attention( 2025-09-07T09:12:30.4349767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4350182Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4350593Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T09:12:30.4351032Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:12:30.4351197Z 2025-09-07T09:12:30.4351300Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4351798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4352275Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4352683Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4353078Z layer_outputs = layer( 2025-09-07T09:12:30.4353458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4353854Z attn_outputs = self.attention( 2025-09-07T09:12:30.4354233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4354651Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4355058Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1324, in forward 2025-09-07T09:12:30.4355595Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T09:12:30.4356178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T09:12:30.4356689Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T09:12:30.4356890Z 2025-09-07T09:12:30.4356985Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4357484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4357973Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4358458Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4358889Z layer_outputs = layer( 2025-09-07T09:12:30.4359264Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4359665Z attn_outputs = self.attention( 2025-09-07T09:12:30.4360055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T09:12:30.4360536Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T09:12:30.4360996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T09:12:30.4361407Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4361544Z 2025-09-07T09:12:30.4361638Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4362171Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4362668Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4363071Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4363470Z layer_outputs = layer( 2025-09-07T09:12:30.4363847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4364334Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4364778Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4365189Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4365586Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4365988Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4366402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T09:12:30.4366823Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4367228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T09:12:30.4367634Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4367765Z 2025-09-07T09:12:30.4367869Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4368368Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4368851Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4369257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4369657Z layer_outputs = layer( 2025-09-07T09:12:30.4370054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4370501Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4370939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4371342Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4371729Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4372125Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4372577Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T09:12:30.4373045Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4373449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T09:12:30.4373858Z hidden_states = self.act_fn(hidden_states) 2025-09-07T09:12:30.4373999Z 2025-09-07T09:12:30.4374105Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4374603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4375093Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4375505Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4375905Z layer_outputs = layer( 2025-09-07T09:12:30.4376317Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4376767Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4377215Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4377620Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4378047Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4378474Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4378881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T09:12:30.4379301Z return self.output(hidden_states) 2025-09-07T09:12:30.4379698Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T09:12:30.4380117Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4380252Z 2025-09-07T09:12:30.4380361Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4380854Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4381344Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4381757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4382155Z layer_outputs = layer( 2025-09-07T09:12:30.4382529Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4382930Z attn_outputs = self.attention( 2025-09-07T09:12:30.4383323Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4383747Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4384167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 540, in forward 2025-09-07T09:12:30.4384586Z query_key_vectors = self.query_key(hidden_states) 2025-09-07T09:12:30.4384742Z 2025-09-07T09:12:30.4384835Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4385337Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4385824Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4386238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4386682Z layer_outputs = layer( 2025-09-07T09:12:30.4387059Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4387516Z attn_outputs = self.attention( 2025-09-07T09:12:30.4387907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4388324Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4388726Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:12:30.4389209Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:12:30.4389707Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 751, in _hash_vectors 2025-09-07T09:12:30.4390212Z rotated_vectors = torch.einsum("bmtd,mdhr->bmhtr", vectors, random_rotations) 2025-09-07T09:12:30.4390461Z 2025-09-07T09:12:30.4390576Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4391087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4391590Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4392020Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4392430Z layer_outputs = layer( 2025-09-07T09:12:30.4392849Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4393254Z attn_outputs = self.attention( 2025-09-07T09:12:30.4393651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4394081Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4394500Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:12:30.4394977Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:12:30.4395463Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T09:12:30.4395950Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T09:12:30.4396154Z 2025-09-07T09:12:30.4396258Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4396760Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4397253Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4397664Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4398062Z layer_outputs = layer( 2025-09-07T09:12:30.4398437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4398844Z attn_outputs = self.attention( 2025-09-07T09:12:30.4399234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4399646Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4400063Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:12:30.4400533Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:12:30.4401024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T09:12:30.4401537Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T09:12:30.4401689Z 2025-09-07T09:12:30.4401789Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4402293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4402788Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4403375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4403786Z layer_outputs = layer( 2025-09-07T09:12:30.4404166Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4404572Z attn_outputs = self.attention( 2025-09-07T09:12:30.4405054Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4405487Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4405898Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:12:30.4406424Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:12:30.4407090Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:12:30.4407624Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T09:12:30.4408077Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T09:12:30.4408580Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T09:12:30.4408799Z 2025-09-07T09:12:30.4408882Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4409120Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4409629Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4410127Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4410544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4410946Z layer_outputs = layer( 2025-09-07T09:12:30.4411325Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4411727Z attn_outputs = self.attention( 2025-09-07T09:12:30.4412120Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4412534Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4412948Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:12:30.4413466Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:12:30.4414079Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:12:30.4414603Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T09:12:30.4415043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T09:12:30.4415483Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T09:12:30.4415686Z 2025-09-07T09:12:30.4415764Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4416011Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4416237Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4416734Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4417229Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4417642Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4418107Z layer_outputs = layer( 2025-09-07T09:12:30.4418482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4418887Z attn_outputs = self.attention( 2025-09-07T09:12:30.4420066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4420500Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4420916Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4421350Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4421775Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T09:12:30.4422323Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4422920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:12:30.4423426Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:12:30.4423631Z 2025-09-07T09:12:30.4423710Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4423919Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4424150Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4424654Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4425149Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4425556Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4425956Z layer_outputs = layer( 2025-09-07T09:12:30.4426335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4426738Z attn_outputs = self.attention( 2025-09-07T09:12:30.4427133Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4427544Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4427957Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4428388Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4428810Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T09:12:30.4429340Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4429888Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:12:30.4430311Z return torch.cat(slices, dim=3) 2025-09-07T09:12:30.4430441Z 2025-09-07T09:12:30.4430516Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4430801Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4431327Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4431820Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4432234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4432630Z layer_outputs = layer( 2025-09-07T09:12:30.4433009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4433487Z attn_outputs = self.attention( 2025-09-07T09:12:30.4433878Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4434301Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4434811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 541, in forward 2025-09-07T09:12:30.4435235Z value_vectors = self.value(hidden_states) 2025-09-07T09:12:30.4435369Z 2025-09-07T09:12:30.4435464Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4435958Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4436443Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4436890Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4437291Z layer_outputs = layer( 2025-09-07T09:12:30.4437660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4438060Z attn_outputs = self.attention( 2025-09-07T09:12:30.4438446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4438864Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4439273Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4439691Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4440108Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T09:12:30.4440619Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4441154Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:12:30.4441648Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:12:30.4441852Z 2025-09-07T09:12:30.4441927Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4442151Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4442650Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4443136Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4443543Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4443933Z layer_outputs = layer( 2025-09-07T09:12:30.4444304Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4444706Z attn_outputs = self.attention( 2025-09-07T09:12:30.4445091Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4445631Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4446043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4446468Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4446887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T09:12:30.4447392Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4447907Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:12:30.4448325Z return torch.cat(slices, dim=3) 2025-09-07T09:12:30.4457081Z 2025-09-07T09:12:30.4457274Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4458014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4458563Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4459038Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4459466Z layer_outputs = layer( 2025-09-07T09:12:30.4459863Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4460322Z attn_outputs = self.attention( 2025-09-07T09:12:30.4460737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4461170Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4461619Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T09:12:30.4462167Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T09:12:30.4462746Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T09:12:30.4463208Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T09:12:30.4463365Z 2025-09-07T09:12:30.4463473Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4463994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4464524Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4464972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4465505Z layer_outputs = layer( 2025-09-07T09:12:30.4465905Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4466318Z attn_outputs = self.attention( 2025-09-07T09:12:30.4466721Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4467154Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4467570Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4468008Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4468438Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T09:12:30.4468926Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:12:30.4469230Z 2025-09-07T09:12:30.4469344Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4469855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4470359Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4470783Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4471188Z layer_outputs = layer( 2025-09-07T09:12:30.4471562Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4471971Z attn_outputs = self.attention( 2025-09-07T09:12:30.4472370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4472832Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4473251Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4473675Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4474099Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T09:12:30.4474627Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4475210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:12:30.4475717Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:12:30.4475922Z 2025-09-07T09:12:30.4476003Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4476249Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4476748Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4477246Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4477662Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4478056Z layer_outputs = layer( 2025-09-07T09:12:30.4478436Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4478840Z attn_outputs = self.attention( 2025-09-07T09:12:30.4479235Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4479657Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4480066Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4480499Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4480920Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T09:12:30.4481367Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:12:30.4481535Z 2025-09-07T09:12:30.4481641Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4482135Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4482625Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4483038Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4483487Z layer_outputs = layer( 2025-09-07T09:12:30.4483897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4484298Z attn_outputs = self.attention( 2025-09-07T09:12:30.4484689Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4485108Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4485522Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4485940Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4486362Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T09:12:30.4486799Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:12:30.4487000Z 2025-09-07T09:12:30.4487089Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4487315Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4487805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4488297Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4488706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4489141Z layer_outputs = layer( 2025-09-07T09:12:30.4489513Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4489910Z attn_outputs = self.attention( 2025-09-07T09:12:30.4490296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4490719Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4491125Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T09:12:30.4491663Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T09:12:30.4492241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T09:12:30.4492770Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T09:12:30.4492978Z 2025-09-07T09:12:30.4493078Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4493585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4494086Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4494501Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4494903Z layer_outputs = layer( 2025-09-07T09:12:30.4495285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4495687Z attn_outputs = self.attention( 2025-09-07T09:12:30.4496087Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T09:12:30.4496555Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T09:12:30.4497029Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T09:12:30.4497448Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4497625Z 2025-09-07T09:12:30.4497766Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4498299Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4498785Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4499195Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4499596Z layer_outputs = layer( 2025-09-07T09:12:30.4499980Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4500429Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4500879Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4501335Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4501737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4502140Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4502560Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T09:12:30.4502998Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4503681Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T09:12:30.4504103Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4504236Z 2025-09-07T09:12:30.4504341Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4504855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4505353Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4505770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4506169Z layer_outputs = layer( 2025-09-07T09:12:30.4506540Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4506994Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4507446Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4507856Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4508243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4508644Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4509055Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T09:12:30.4509487Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4509889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T09:12:30.4510299Z hidden_states = self.act_fn(hidden_states) 2025-09-07T09:12:30.4510443Z 2025-09-07T09:12:30.4510542Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4511046Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4511539Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4511954Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4512469Z layer_outputs = layer( 2025-09-07T09:12:30.4512847Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4513299Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4513749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4514154Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4514542Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4514935Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4515347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T09:12:30.4515778Z return self.output(hidden_states) 2025-09-07T09:12:30.4516296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T09:12:30.4516723Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4516867Z 2025-09-07T09:12:30.4516965Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4517465Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4517958Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4518402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4518808Z layer_outputs = layer( 2025-09-07T09:12:30.4519187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4519594Z attn_outputs = self.attention( 2025-09-07T09:12:30.4519989Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4520404Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4520819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1198, in forward 2025-09-07T09:12:30.4521228Z query_vectors = self.query(hidden_states) 2025-09-07T09:12:30.4521362Z 2025-09-07T09:12:30.4521468Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4521968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4522451Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4522860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4523257Z layer_outputs = layer( 2025-09-07T09:12:30.4523630Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4524033Z attn_outputs = self.attention( 2025-09-07T09:12:30.4524417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4524831Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4525242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1199, in forward 2025-09-07T09:12:30.4525647Z key_vectors = self.key(hidden_states) 2025-09-07T09:12:30.4525775Z 2025-09-07T09:12:30.4525856Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4526051Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4526314Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4526846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4527337Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4527736Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4528131Z layer_outputs = layer( 2025-09-07T09:12:30.4528506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4528906Z attn_outputs = self.attention( 2025-09-07T09:12:30.4529288Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4529696Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4530140Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T09:12:30.4530661Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4531197Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:12:30.4531694Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:12:30.4531899Z 2025-09-07T09:12:30.4532050Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4532287Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4532508Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4533008Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4533501Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4533906Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4534303Z layer_outputs = layer( 2025-09-07T09:12:30.4534674Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4535064Z attn_outputs = self.attention( 2025-09-07T09:12:30.4535449Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4535864Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4536265Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T09:12:30.4536766Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4537292Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:12:30.4537779Z return torch.cat(slices, dim=3) 2025-09-07T09:12:30.4537901Z 2025-09-07T09:12:30.4538097Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4538590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4539070Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4539467Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4539857Z layer_outputs = layer( 2025-09-07T09:12:30.4540225Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4540696Z attn_outputs = self.attention( 2025-09-07T09:12:30.4541081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4541538Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4541939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T09:12:30.4542407Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:12:30.4542608Z 2025-09-07T09:12:30.4542700Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4543190Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4543673Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4544114Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4544516Z layer_outputs = layer( 2025-09-07T09:12:30.4544882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4545276Z attn_outputs = self.attention( 2025-09-07T09:12:30.4545658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4546071Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4546508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T09:12:30.4546977Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:12:30.4547179Z 2025-09-07T09:12:30.4547274Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4547764Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4548256Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4548661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4549049Z layer_outputs = layer( 2025-09-07T09:12:30.4549411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4549806Z attn_outputs = self.attention( 2025-09-07T09:12:30.4550187Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4550592Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4550994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1200, in forward 2025-09-07T09:12:30.4551398Z value_vectors = self.value(hidden_states) 2025-09-07T09:12:30.4551529Z 2025-09-07T09:12:30.4551605Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4551801Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4552012Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4552510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4552994Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4553398Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4553794Z layer_outputs = layer( 2025-09-07T09:12:30.4554157Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4554596Z attn_outputs = self.attention( 2025-09-07T09:12:30.4555010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4555419Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4555819Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1266, in forward 2025-09-07T09:12:30.4556335Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4556872Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:12:30.4557360Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:12:30.4557561Z 2025-09-07T09:12:30.4557636Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4557822Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4558066Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4558566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4559043Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4559443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4559830Z layer_outputs = layer( 2025-09-07T09:12:30.4560247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4560644Z attn_outputs = self.attention( 2025-09-07T09:12:30.4561022Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4561432Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4561828Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T09:12:30.4562267Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:12:30.4562432Z 2025-09-07T09:12:30.4562527Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4563014Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4563493Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4563900Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4564298Z layer_outputs = layer( 2025-09-07T09:12:30.4564669Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4565069Z attn_outputs = self.attention( 2025-09-07T09:12:30.4565457Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4565864Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4566267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T09:12:30.4566699Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:12:30.4566867Z 2025-09-07T09:12:30.4566965Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4567454Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4567932Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4568335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4568799Z layer_outputs = layer( 2025-09-07T09:12:30.4569165Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4569557Z attn_outputs = self.attention( 2025-09-07T09:12:30.4569935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4570345Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4570749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1324, in forward 2025-09-07T09:12:30.4571281Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T09:12:30.4571889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T09:12:30.4572409Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T09:12:30.4572605Z 2025-09-07T09:12:30.4572698Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4573192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4573675Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4574103Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4574498Z layer_outputs = layer( 2025-09-07T09:12:30.4574861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4575253Z attn_outputs = self.attention( 2025-09-07T09:12:30.4575634Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T09:12:30.4576094Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T09:12:30.4576557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T09:12:30.4576960Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4577090Z 2025-09-07T09:12:30.4577186Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4577720Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4578200Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4578603Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4578994Z layer_outputs = layer( 2025-09-07T09:12:30.4579359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4579796Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4580228Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4580624Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4581009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4581393Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4581796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T09:12:30.4582213Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4582658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T09:12:30.4583098Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4583223Z 2025-09-07T09:12:30.4583319Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4583812Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4584292Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4584696Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4585093Z layer_outputs = layer( 2025-09-07T09:12:30.4585460Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4585935Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4586381Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4586784Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4587168Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4587558Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4587994Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T09:12:30.4588424Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4588822Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T09:12:30.4589230Z hidden_states = self.act_fn(hidden_states) 2025-09-07T09:12:30.4589359Z 2025-09-07T09:12:30.4589459Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4589942Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4590417Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4590813Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4591200Z layer_outputs = layer( 2025-09-07T09:12:30.4591566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4591998Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4592427Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4592827Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4593201Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4593587Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4593981Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T09:12:30.4594395Z return self.output(hidden_states) 2025-09-07T09:12:30.4594777Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T09:12:30.4595187Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4595318Z 2025-09-07T09:12:30.4595412Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4595904Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4596431Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4596868Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4597261Z layer_outputs = layer( 2025-09-07T09:12:30.4597626Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4598020Z attn_outputs = self.attention( 2025-09-07T09:12:30.4598402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4598813Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4599220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 540, in forward 2025-09-07T09:12:30.4599632Z query_key_vectors = self.query_key(hidden_states) 2025-09-07T09:12:30.4599813Z 2025-09-07T09:12:30.4599905Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4600402Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4600879Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4601279Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4601678Z layer_outputs = layer( 2025-09-07T09:12:30.4602081Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4602487Z attn_outputs = self.attention( 2025-09-07T09:12:30.4602874Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4603438Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4603846Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:12:30.4604313Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:12:30.4604802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 751, in _hash_vectors 2025-09-07T09:12:30.4605291Z rotated_vectors = torch.einsum("bmtd,mdhr->bmhtr", vectors, random_rotations) 2025-09-07T09:12:30.4605497Z 2025-09-07T09:12:30.4605601Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4606089Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4606574Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4606986Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4607379Z layer_outputs = layer( 2025-09-07T09:12:30.4607747Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4608137Z attn_outputs = self.attention( 2025-09-07T09:12:30.4608526Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4608939Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4609359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:12:30.4609819Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:12:30.4610305Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T09:12:30.4611225Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T09:12:30.4611429Z 2025-09-07T09:12:30.4611525Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4612016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4612499Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4612901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4613295Z layer_outputs = layer( 2025-09-07T09:12:30.4613661Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4614054Z attn_outputs = self.attention( 2025-09-07T09:12:30.4614499Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4614924Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4615334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:12:30.4615800Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:12:30.4616285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T09:12:30.4616760Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T09:12:30.4616914Z 2025-09-07T09:12:30.4617013Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4617510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4618039Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4618451Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4618840Z layer_outputs = layer( 2025-09-07T09:12:30.4619210Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4619605Z attn_outputs = self.attention( 2025-09-07T09:12:30.4619993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4620404Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4620804Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:12:30.4621326Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:12:30.4621935Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:12:30.4622454Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T09:12:30.4622892Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T09:12:30.4623403Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T09:12:30.4623614Z 2025-09-07T09:12:30.4623690Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4623912Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4624405Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4624944Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4625349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4625781Z layer_outputs = layer( 2025-09-07T09:12:30.4626152Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4626548Z attn_outputs = self.attention( 2025-09-07T09:12:30.4626931Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4627378Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4627780Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:12:30.4628290Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:12:30.4628933Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:12:30.4629456Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T09:12:30.4629894Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T09:12:30.4630330Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T09:12:30.4630475Z 2025-09-07T09:12:30.4630549Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4630781Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4630992Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4631487Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4631979Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4632388Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4632784Z layer_outputs = layer( 2025-09-07T09:12:30.4633149Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4633543Z attn_outputs = self.attention( 2025-09-07T09:12:30.4633932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4634343Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4634749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4635165Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4635587Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T09:12:30.4636120Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4636680Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:12:30.4637169Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:12:30.4637371Z 2025-09-07T09:12:30.4637443Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4637632Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4637848Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4638336Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4638820Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4639267Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4639694Z layer_outputs = layer( 2025-09-07T09:12:30.4640063Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4640458Z attn_outputs = self.attention( 2025-09-07T09:12:30.4640838Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4641247Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4641649Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4642065Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4642520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T09:12:30.4643048Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4643590Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:12:30.4644018Z return torch.cat(slices, dim=3) 2025-09-07T09:12:30.4644139Z 2025-09-07T09:12:30.4644215Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4644433Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4644968Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4645452Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4645855Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4646247Z layer_outputs = layer( 2025-09-07T09:12:30.4646609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4646996Z attn_outputs = self.attention( 2025-09-07T09:12:30.4647376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4647786Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4648192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 541, in forward 2025-09-07T09:12:30.4648594Z value_vectors = self.value(hidden_states) 2025-09-07T09:12:30.4648727Z 2025-09-07T09:12:30.4648818Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4649311Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4649803Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4650213Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4650607Z layer_outputs = layer( 2025-09-07T09:12:30.4650977Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4651373Z attn_outputs = self.attention( 2025-09-07T09:12:30.4651762Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4652173Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4652575Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4653061Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4653514Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T09:12:30.4654021Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4654546Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:12:30.4655029Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:12:30.4655236Z 2025-09-07T09:12:30.4655311Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4655541Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4656034Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4656564Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4656972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4657365Z layer_outputs = layer( 2025-09-07T09:12:30.4657796Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4658192Z attn_outputs = self.attention( 2025-09-07T09:12:30.4658573Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4659070Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4659484Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4659912Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4660332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T09:12:30.4660835Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4661346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:12:30.4661764Z return torch.cat(slices, dim=3) 2025-09-07T09:12:30.4661889Z 2025-09-07T09:12:30.4661983Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4662476Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4662959Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4663358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4663753Z layer_outputs = layer( 2025-09-07T09:12:30.4664123Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4664523Z attn_outputs = self.attention( 2025-09-07T09:12:30.4664903Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4665314Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4665712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T09:12:30.4666244Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T09:12:30.4666815Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T09:12:30.4667319Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T09:12:30.4667521Z 2025-09-07T09:12:30.4667613Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4668101Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4668577Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4668974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4669357Z layer_outputs = layer( 2025-09-07T09:12:30.4669724Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4670114Z attn_outputs = self.attention( 2025-09-07T09:12:30.4670532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4670952Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4671350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4671769Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4672184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T09:12:30.4672656Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:12:30.4672859Z 2025-09-07T09:12:30.4672993Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4673485Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4673968Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4674375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4674766Z layer_outputs = layer( 2025-09-07T09:12:30.4675128Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4675519Z attn_outputs = self.attention( 2025-09-07T09:12:30.4675899Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4676316Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4676716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4677127Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4677538Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T09:12:30.4678052Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4678585Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:12:30.4679068Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:12:30.4679269Z 2025-09-07T09:12:30.4679344Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4679563Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4680052Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4680535Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4680936Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4681402Z layer_outputs = layer( 2025-09-07T09:12:30.4681763Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4682157Z attn_outputs = self.attention( 2025-09-07T09:12:30.4682539Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4682947Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4683345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4683761Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4684170Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T09:12:30.4684647Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:12:30.4684815Z 2025-09-07T09:12:30.4684912Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4685392Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4685887Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4686295Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4686692Z layer_outputs = layer( 2025-09-07T09:12:30.4687096Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4687497Z attn_outputs = self.attention( 2025-09-07T09:12:30.4687889Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4688312Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4688725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4689147Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4689557Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T09:12:30.4689991Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:12:30.4690158Z 2025-09-07T09:12:30.4690233Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4690454Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4690945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4691426Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4691834Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4692228Z layer_outputs = layer( 2025-09-07T09:12:30.4692597Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4692991Z attn_outputs = self.attention( 2025-09-07T09:12:30.4693376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4693789Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4694192Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T09:12:30.4694726Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T09:12:30.4695349Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T09:12:30.4695901Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T09:12:30.4696103Z 2025-09-07T09:12:30.4696197Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4696703Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4697189Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4697600Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4698057Z layer_outputs = layer( 2025-09-07T09:12:30.4698443Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4698890Z attn_outputs = self.attention( 2025-09-07T09:12:30.4699287Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T09:12:30.4699753Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T09:12:30.4700222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T09:12:30.4700634Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4700769Z 2025-09-07T09:12:30.4700908Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4701409Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4701895Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4702313Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4702715Z layer_outputs = layer( 2025-09-07T09:12:30.4703227Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1601, in forward 2025-09-07T09:12:30.4703653Z attn_output = prev_attn_output + attn_output 2025-09-07T09:12:30.4703794Z 2025-09-07T09:12:30.4703888Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4704386Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4704873Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4705283Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4705676Z layer_outputs = layer( 2025-09-07T09:12:30.4706048Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4706504Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4706952Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4707362Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4707757Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4708157Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4708583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T09:12:30.4709023Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4709444Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T09:12:30.4709989Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4710132Z 2025-09-07T09:12:30.4710234Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4710737Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4711224Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4711644Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4712035Z layer_outputs = layer( 2025-09-07T09:12:30.4712413Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4712863Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4713370Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4713786Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4714172Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4714568Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4714978Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T09:12:30.4715405Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4715860Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T09:12:30.4716270Z hidden_states = self.act_fn(hidden_states) 2025-09-07T09:12:30.4716409Z 2025-09-07T09:12:30.4716507Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4717013Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4717511Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4717919Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4718313Z layer_outputs = layer( 2025-09-07T09:12:30.4718691Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4719142Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4719583Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4719987Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4720371Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4720763Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4721173Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T09:12:30.4721591Z return self.output(hidden_states) 2025-09-07T09:12:30.4721974Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T09:12:30.4722380Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4722385Z 2025-09-07T09:12:30.4722487Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4722833Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4722918Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4723247Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4723344Z layer_outputs = layer( 2025-09-07T09:12:30.4723620Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4723733Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4723737Z 2025-09-07T09:12:30.4723838Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4724181Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4724266Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4724532Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4724631Z layer_outputs = layer( 2025-09-07T09:12:30.4724911Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4724981Z attn_outputs = self.attention( 2025-09-07T09:12:30.4725258Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4725346Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4725645Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1198, in forward 2025-09-07T09:12:30.4725733Z query_vectors = self.query(hidden_states) 2025-09-07T09:12:30.4725737Z 2025-09-07T09:12:30.4725829Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4726178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4726259Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4726534Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4726595Z layer_outputs = layer( 2025-09-07T09:12:30.4726861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4726937Z attn_outputs = self.attention( 2025-09-07T09:12:30.4727206Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4727294Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4727563Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1199, in forward 2025-09-07T09:12:30.4727637Z key_vectors = self.key(hidden_states) 2025-09-07T09:12:30.4727642Z 2025-09-07T09:12:30.4727726Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4727796Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4727902Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4728242Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4728332Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4728606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4728671Z layer_outputs = layer( 2025-09-07T09:12:30.4728945Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4729015Z attn_outputs = self.attention( 2025-09-07T09:12:30.4729335Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4729526Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4729798Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T09:12:30.4729991Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4730281Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:12:30.4730454Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:12:30.4730458Z 2025-09-07T09:12:30.4730535Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4730627Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4730726Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4731107Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4731199Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4731468Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4731539Z layer_outputs = layer( 2025-09-07T09:12:30.4731805Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4731906Z attn_outputs = self.attention( 2025-09-07T09:12:30.4732184Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4732269Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4732544Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1265, in forward 2025-09-07T09:12:30.4732719Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4733009Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:12:30.4733080Z return torch.cat(slices, dim=3) 2025-09-07T09:12:30.4733083Z 2025-09-07T09:12:30.4733179Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4733530Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4733611Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4733887Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4733953Z layer_outputs = layer( 2025-09-07T09:12:30.4734220Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4734297Z attn_outputs = self.attention( 2025-09-07T09:12:30.4734564Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4734652Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4734917Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T09:12:30.4735073Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:12:30.4735076Z 2025-09-07T09:12:30.4735169Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4735510Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4735661Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4735930Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4735997Z layer_outputs = layer( 2025-09-07T09:12:30.4736263Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4736336Z attn_outputs = self.attention( 2025-09-07T09:12:30.4736606Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4736687Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4736960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1272, in forward 2025-09-07T09:12:30.4737134Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:12:30.4737139Z 2025-09-07T09:12:30.4737238Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4737579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4737737Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4738007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4738067Z layer_outputs = layer( 2025-09-07T09:12:30.4738376Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4738442Z attn_outputs = self.attention( 2025-09-07T09:12:30.4738717Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4738801Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4739069Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1200, in forward 2025-09-07T09:12:30.4739157Z value_vectors = self.value(hidden_states) 2025-09-07T09:12:30.4739161Z 2025-09-07T09:12:30.4739237Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4739314Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4739406Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4739749Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4739833Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4740098Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4740165Z layer_outputs = layer( 2025-09-07T09:12:30.4740432Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4740502Z attn_outputs = self.attention( 2025-09-07T09:12:30.4740767Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4740844Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4741119Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1266, in forward 2025-09-07T09:12:30.4741313Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4741607Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:12:30.4741783Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:12:30.4741816Z 2025-09-07T09:12:30.4741891Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4741960Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4742052Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4742417Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4742494Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4742768Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4742828Z layer_outputs = layer( 2025-09-07T09:12:30.4743104Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4743183Z attn_outputs = self.attention( 2025-09-07T09:12:30.4743482Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4743585Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4743861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T09:12:30.4743983Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:12:30.4743987Z 2025-09-07T09:12:30.4744076Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4744460Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4744541Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4744808Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4744877Z layer_outputs = layer( 2025-09-07T09:12:30.4745141Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4745205Z attn_outputs = self.attention( 2025-09-07T09:12:30.4745473Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4745550Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4745820Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1308, in forward 2025-09-07T09:12:30.4745922Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:12:30.4745925Z 2025-09-07T09:12:30.4746017Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4746355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4746435Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4746701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4746759Z layer_outputs = layer( 2025-09-07T09:12:30.4747028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4747091Z attn_outputs = self.attention( 2025-09-07T09:12:30.4747358Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4747438Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4747702Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1324, in forward 2025-09-07T09:12:30.4747950Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T09:12:30.4748303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T09:12:30.4748444Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T09:12:30.4748447Z 2025-09-07T09:12:30.4748544Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4748893Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4748976Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4749243Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4749312Z layer_outputs = layer( 2025-09-07T09:12:30.4749617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4749694Z attn_outputs = self.attention( 2025-09-07T09:12:30.4749960Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T09:12:30.4750093Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T09:12:30.4750366Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T09:12:30.4750477Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4750480Z 2025-09-07T09:12:30.4750583Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4750926Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4751017Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4751285Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4751346Z layer_outputs = layer( 2025-09-07T09:12:30.4751621Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4751740Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4752016Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4752089Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4752350Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4752429Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4752718Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T09:12:30.4752802Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4753074Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T09:12:30.4753163Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4753166Z 2025-09-07T09:12:30.4753260Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4753602Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4753690Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4753956Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4754066Z layer_outputs = layer( 2025-09-07T09:12:30.4754334Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4754482Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4754755Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4754830Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4755097Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4755163Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4755456Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T09:12:30.4755531Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4755832Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T09:12:30.4755919Z hidden_states = self.act_fn(hidden_states) 2025-09-07T09:12:30.4755922Z 2025-09-07T09:12:30.4756012Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4756355Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4756437Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4761811Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4761913Z layer_outputs = layer( 2025-09-07T09:12:30.4762231Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4762371Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4762658Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4762735Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4763010Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4763078Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4763375Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T09:12:30.4763443Z return self.output(hidden_states) 2025-09-07T09:12:30.4763725Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T09:12:30.4763808Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4763814Z 2025-09-07T09:12:30.4763926Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4764322Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4764414Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4764712Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4764780Z layer_outputs = layer( 2025-09-07T09:12:30.4765067Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4765142Z attn_outputs = self.attention( 2025-09-07T09:12:30.4765411Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4765502Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4765823Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 540, in forward 2025-09-07T09:12:30.4765954Z query_key_vectors = self.query_key(hidden_states) 2025-09-07T09:12:30.4765958Z 2025-09-07T09:12:30.4766057Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4766423Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4766515Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4766793Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4766862Z layer_outputs = layer( 2025-09-07T09:12:30.4767132Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4767249Z attn_outputs = self.attention( 2025-09-07T09:12:30.4767520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4767607Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4767881Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:12:30.4768024Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:12:30.4768347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 751, in _hash_vectors 2025-09-07T09:12:30.4768498Z rotated_vectors = torch.einsum("bmtd,mdhr->bmhtr", vectors, random_rotations) 2025-09-07T09:12:30.4768501Z 2025-09-07T09:12:30.4768600Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4768946Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4769029Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4769303Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4769367Z layer_outputs = layer( 2025-09-07T09:12:30.4769641Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4769711Z attn_outputs = self.attention( 2025-09-07T09:12:30.4769983Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4770070Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4770346Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:12:30.4770492Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:12:30.4770779Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 754, in _hash_vectors 2025-09-07T09:12:30.4770918Z rotated_vectors = torch.cat([rotated_vectors, -rotated_vectors], dim=-1) 2025-09-07T09:12:30.4770922Z 2025-09-07T09:12:30.4771020Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4771367Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4771450Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4771716Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4771780Z layer_outputs = layer( 2025-09-07T09:12:30.4772083Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4772183Z attn_outputs = self.attention( 2025-09-07T09:12:30.4772452Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4772531Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4772802Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 578, in forward 2025-09-07T09:12:30.4772934Z buckets = self._hash_vectors(query_key_vectors, num_hashes, attention_mask) 2025-09-07T09:12:30.4773218Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 755, in _hash_vectors 2025-09-07T09:12:30.4773304Z buckets = torch.argmax(rotated_vectors, dim=-1) 2025-09-07T09:12:30.4773307Z 2025-09-07T09:12:30.4773448Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4773797Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4773877Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4774146Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4774207Z layer_outputs = layer( 2025-09-07T09:12:30.4774508Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4774572Z attn_outputs = self.attention( 2025-09-07T09:12:30.4774835Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4774920Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4775189Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:12:30.4775377Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:12:30.4775739Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:12:30.4775839Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T09:12:30.4776131Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 182, in _stable_argsort 2025-09-07T09:12:30.4776276Z scaled_vector = vector.shape[dim] * vector + (scale_offset % vector.shape[dim]) 2025-09-07T09:12:30.4776279Z 2025-09-07T09:12:30.4776357Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4776449Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4776801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4776882Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4777148Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4777211Z layer_outputs = layer( 2025-09-07T09:12:30.4777478Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4777555Z attn_outputs = self.attention( 2025-09-07T09:12:30.4777897Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4777979Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4778249Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 587, in forward 2025-09-07T09:12:30.4778501Z sorted_bucket_idx, undo_sorted_bucket_idx = self._get_sorted_bucket_idx_and_undo_sorted_bucket_idx( 2025-09-07T09:12:30.4778861Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 796, in _get_sorted_bucket_idx_and_undo_sorted_bucket_idx 2025-09-07T09:12:30.4778954Z sorted_bucket_idx = _stable_argsort(buckets, dim=-1) 2025-09-07T09:12:30.4779250Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 183, in _stable_argsort 2025-09-07T09:12:30.4779336Z return torch.argsort(scaled_vector, dim=dim) 2025-09-07T09:12:30.4779339Z 2025-09-07T09:12:30.4779418Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4779487Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4779578Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4779964Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4780047Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4780315Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4780375Z layer_outputs = layer( 2025-09-07T09:12:30.4780643Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4780746Z attn_outputs = self.attention( 2025-09-07T09:12:30.4781015Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4781099Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4781373Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4781466Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4781741Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T09:12:30.4781947Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4782241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:12:30.4782390Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:12:30.4782393Z 2025-09-07T09:12:30.4782471Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4782539Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4782633Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4782982Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4783063Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4783339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4783402Z layer_outputs = layer( 2025-09-07T09:12:30.4783676Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4783744Z attn_outputs = self.attention( 2025-09-07T09:12:30.4784007Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4784090Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4784353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4784509Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4784773Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 859, in _attend 2025-09-07T09:12:30.4784972Z key_value_bucket_idx = self._look_adjacent(query_bucket_idx, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4785257Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:12:30.4785324Z return torch.cat(slices, dim=3) 2025-09-07T09:12:30.4785327Z 2025-09-07T09:12:30.4785400Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4785488Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4785865Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4785946Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4786212Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4786273Z layer_outputs = layer( 2025-09-07T09:12:30.4786537Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4786605Z attn_outputs = self.attention( 2025-09-07T09:12:30.4786901Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4786981Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4787248Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 541, in forward 2025-09-07T09:12:30.4787321Z value_vectors = self.value(hidden_states) 2025-09-07T09:12:30.4787329Z 2025-09-07T09:12:30.4787422Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4787759Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4787839Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4788102Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4788164Z layer_outputs = layer( 2025-09-07T09:12:30.4788437Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4788501Z attn_outputs = self.attention( 2025-09-07T09:12:30.4788770Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4788848Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4789113Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4789202Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4789464Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T09:12:30.4789639Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4789922Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:12:30.4790058Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:12:30.4790062Z 2025-09-07T09:12:30.4790129Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4790254Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4790594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4790704Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4790972Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4791031Z layer_outputs = layer( 2025-09-07T09:12:30.4791296Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4791360Z attn_outputs = self.attention( 2025-09-07T09:12:30.4791623Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4791701Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4791999Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4792091Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4792353Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 844, in _attend 2025-09-07T09:12:30.4792519Z key_vectors = self._look_adjacent(key_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4792801Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 400, in _look_adjacent 2025-09-07T09:12:30.4792895Z return torch.cat(slices, dim=3) 2025-09-07T09:12:30.4792899Z 2025-09-07T09:12:30.4792992Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4793329Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4793409Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4793673Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4793731Z layer_outputs = layer( 2025-09-07T09:12:30.4793996Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4794058Z attn_outputs = self.attention( 2025-09-07T09:12:30.4794324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4794399Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4794660Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 595, in forward 2025-09-07T09:12:30.4794857Z query_key_vectors = self._gather_by_expansion(query_key_vectors, sorted_bucket_idx_per_hash, num_hashes) 2025-09-07T09:12:30.4795167Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1099, in _gather_by_expansion 2025-09-07T09:12:30.4795254Z return torch.gather(vectors, 2, expanded_idxs) 2025-09-07T09:12:30.4795257Z 2025-09-07T09:12:30.4795345Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4795682Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4795758Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4796024Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4796080Z layer_outputs = layer( 2025-09-07T09:12:30.4796347Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4796558Z attn_outputs = self.attention( 2025-09-07T09:12:30.4796873Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4796957Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4797222Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4797312Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4797589Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 849, in _attend 2025-09-07T09:12:30.4797738Z query_key_dots = torch.matmul(query_vectors, key_vectors.transpose(-1, -2)) 2025-09-07T09:12:30.4797741Z 2025-09-07T09:12:30.4797846Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4798238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4798335Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4798618Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4798685Z layer_outputs = layer( 2025-09-07T09:12:30.4798963Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4799034Z attn_outputs = self.attention( 2025-09-07T09:12:30.4799340Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4799431Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4799706Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4799812Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4800088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 845, in _attend 2025-09-07T09:12:30.4800288Z value_vectors = self._look_adjacent(value_vectors, self.num_chunks_before, self.num_chunks_after) 2025-09-07T09:12:30.4800579Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 399, in _look_adjacent 2025-09-07T09:12:30.4800728Z slices.append(torch.cat([vectors[:, :, i:, ...], vectors[:, :, :i, ...]], dim=2)) 2025-09-07T09:12:30.4800732Z 2025-09-07T09:12:30.4800808Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4800909Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4801259Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4801344Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4801613Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4801683Z layer_outputs = layer( 2025-09-07T09:12:30.4801951Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4802024Z attn_outputs = self.attention( 2025-09-07T09:12:30.4802293Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4802379Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4802655Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4802786Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4803241Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T09:12:30.4803417Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:12:30.4803421Z 2025-09-07T09:12:30.4803521Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4803867Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4803946Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4804226Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4804290Z layer_outputs = layer( 2025-09-07T09:12:30.4804609Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4804678Z attn_outputs = self.attention( 2025-09-07T09:12:30.4804953Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4805034Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4805301Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 637, in forward 2025-09-07T09:12:30.4805395Z out_vectors, logits, attention_probs = self._attend( 2025-09-07T09:12:30.4805714Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 931, in _attend 2025-09-07T09:12:30.4805822Z out_vectors = torch.matmul(attention_probs, value_vectors) 2025-09-07T09:12:30.4805825Z 2025-09-07T09:12:30.4805898Z cudagraph partition due to non gpu ops 2025-09-07T09:12:30.4805991Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4806345Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4806424Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4806697Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4806757Z layer_outputs = layer( 2025-09-07T09:12:30.4807028Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4807094Z attn_outputs = self.attention( 2025-09-07T09:12:30.4807359Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1417, in forward 2025-09-07T09:12:30.4807440Z self_attention_outputs = self.self_attention( 2025-09-07T09:12:30.4807705Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 692, in forward 2025-09-07T09:12:30.4807917Z out_vectors = self._merge_hidden_size_dims(out_vectors, self.num_attention_heads, self.attention_head_size) 2025-09-07T09:12:30.4808234Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 415, in _merge_hidden_size_dims 2025-09-07T09:12:30.4808375Z return torch.reshape(x, (x.size()[0], -1, num_attn_heads * attn_head_size)) 2025-09-07T09:12:30.4808384Z 2025-09-07T09:12:30.4808475Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4808821Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4808910Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4809178Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4809292Z layer_outputs = layer( 2025-09-07T09:12:30.4809594Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1587, in forward 2025-09-07T09:12:30.4809662Z attn_outputs = self.attention( 2025-09-07T09:12:30.4809932Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1457, in forward 2025-09-07T09:12:30.4810066Z attention_output = self.output(self_attention_outputs.hidden_states) 2025-09-07T09:12:30.4810339Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1371, in forward 2025-09-07T09:12:30.4810416Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4810419Z 2025-09-07T09:12:30.4810512Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4810884Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4810967Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4811238Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4811297Z layer_outputs = layer( 2025-09-07T09:12:30.4811566Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4811690Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4811993Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4812066Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4812332Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4812410Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4812701Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T09:12:30.4812776Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4813043Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1479, in forward 2025-09-07T09:12:30.4813114Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4813117Z 2025-09-07T09:12:30.4813214Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4813554Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4813635Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4813910Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4813977Z layer_outputs = layer( 2025-09-07T09:12:30.4814239Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4814349Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4814617Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4814683Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4814944Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4815012Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4815302Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1518, in forward_chunk 2025-09-07T09:12:30.4815416Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4815715Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1481, in forward 2025-09-07T09:12:30.4815794Z hidden_states = self.act_fn(hidden_states) 2025-09-07T09:12:30.4815798Z 2025-09-07T09:12:30.4815889Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4816233Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4816311Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4816576Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4816639Z layer_outputs = layer( 2025-09-07T09:12:30.4816939Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4817057Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4817324Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1509, in forward 2025-09-07T09:12:30.4817389Z return apply_chunking_to_forward( 2025-09-07T09:12:30.4817651Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:30.4817763Z return forward_fn(*input_tensors) 2025-09-07T09:12:30.4818088Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1519, in forward_chunk 2025-09-07T09:12:30.4818155Z return self.output(hidden_states) 2025-09-07T09:12:30.4818425Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1493, in forward 2025-09-07T09:12:30.4818502Z hidden_states = self.dense(hidden_states) 2025-09-07T09:12:30.4818506Z 2025-09-07T09:12:30.4818596Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4818940Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4819018Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4819286Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1722, in forward 2025-09-07T09:12:30.4819347Z layer_outputs = layer( 2025-09-07T09:12:30.4819615Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1612, in forward 2025-09-07T09:12:30.4819725Z hidden_states = hidden_states + self.feed_forward(attn_output) 2025-09-07T09:12:30.4819728Z 2025-09-07T09:12:30.4819815Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:30.4820156Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1841, in torch_dynamo_resume_in_forward_at_1837 2025-09-07T09:12:30.4820235Z hidden_states = _ReversibleFunction.apply( 2025-09-07T09:12:30.4820506Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1753, in forward 2025-09-07T09:12:30.4820609Z return torch.cat([attn_output, hidden_states], dim=-1) 2025-09-07T09:12:30.4820612Z 2025-09-07T09:12:44.9421252Z cudagraph partition due to non gpu ops. Found from : 2025-09-07T09:12:44.9421913Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 2495, in torch_dynamo_resume_in_forward_at_2481 2025-09-07T09:12:44.9422425Z logits = self.lm_head(sequence_output) 2025-09-07T09:12:44.9422882Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1886, in forward 2025-09-07T09:12:44.9423855Z return apply_chunking_to_forward(self.forward_chunk, self.chunk_size_lm_head, self.seq_len_dim, hidden_states) 2025-09-07T09:12:44.9424520Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/pytorch_utils.py", line 251, in apply_chunking_to_forward 2025-09-07T09:12:44.9424929Z return forward_fn(*input_tensors) 2025-09-07T09:12:44.9425363Z File "/opt/conda/envs/py_3.9/lib/python3.9/site-packages/transformers/models/reformer/modeling_reformer.py", line 1889, in forward_chunk 2025-09-07T09:12:44.9425808Z hidden_states = self.decoder(hidden_states) 2025-09-07T09:12:44.9425955Z 2025-09-07T09:12:45.7496515Z 2025-09-07T09:12:46.1707240Z running benchmark: 0% 0/30 [00:00> $GITHUB_ENV 2025-09-07T09:49:41.4824575Z echo "DEVICE_TYPE=$DEVICE_TYPE" >> $GITHUB_ENV 2025-09-07T09:49:41.4836000Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:49:41.4836244Z env: 2025-09-07T09:49:41.4836396Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:41.4836706Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:41.4837018Z ##[endgroup] 2025-09-07T09:49:41.4871500Z + [[ -n '' ]] 2025-09-07T09:49:41.4871842Z + python3 -mpip install boto3==1.35.33 psutil==7.0.0 pynvml==12.0.0 2025-09-07T09:49:41.6632924Z Defaulting to user installation because normal site-packages is not writeable 2025-09-07T09:49:42.4325850Z Collecting boto3==1.35.33 2025-09-07T09:49:42.4452287Z Downloading boto3-1.35.33-py3-none-any.whl (139 kB) 2025-09-07T09:49:42.6601311Z Collecting psutil==7.0.0 2025-09-07T09:49:42.6631206Z Downloading psutil-7.0.0-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (277 kB) 2025-09-07T09:49:42.6915955Z Collecting pynvml==12.0.0 2025-09-07T09:49:42.6940260Z Downloading pynvml-12.0.0-py3-none-any.whl (26 kB) 2025-09-07T09:49:42.7335348Z Collecting s3transfer<0.11.0,>=0.10.0 2025-09-07T09:49:42.7359916Z Downloading s3transfer-0.10.4-py3-none-any.whl (83 kB) 2025-09-07T09:49:42.7407456Z Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /usr/lib/python3.9/site-packages (from boto3==1.35.33) (0.10.0) 2025-09-07T09:49:43.5231156Z Collecting botocore<1.36.0,>=1.35.33 2025-09-07T09:49:43.5263263Z Downloading botocore-1.35.99-py3-none-any.whl (13.3 MB) 2025-09-07T09:49:43.6645451Z Collecting nvidia-ml-py<13.0.0a0,>=12.0.0 2025-09-07T09:49:43.6669401Z Downloading nvidia_ml_py-12.575.51-py3-none-any.whl (47 kB) 2025-09-07T09:49:43.6734507Z Requirement already satisfied: python-dateutil<3.0.0,>=2.1 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (2.8.1) 2025-09-07T09:49:43.6741787Z Requirement already satisfied: urllib3<1.27,>=1.25.4 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.25.10) 2025-09-07T09:49:43.8125725Z Requirement already satisfied: six>=1.5 in /usr/lib/python3.9/site-packages (from python-dateutil<3.0.0,>=2.1->botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.15.0) 2025-09-07T09:49:43.9056856Z Installing collected packages: botocore, s3transfer, nvidia-ml-py, pynvml, psutil, boto3 2025-09-07T09:49:44.3109477Z Attempting uninstall: nvidia-ml-py 2025-09-07T09:49:44.3110643Z Found existing installation: nvidia-ml-py 11.525.84 2025-09-07T09:49:44.3121411Z Uninstalling nvidia-ml-py-11.525.84: 2025-09-07T09:49:44.3303867Z Successfully uninstalled nvidia-ml-py-11.525.84 2025-09-07T09:49:44.3733026Z Attempting uninstall: psutil 2025-09-07T09:49:44.3733650Z Found existing installation: psutil 5.9.8 2025-09-07T09:49:44.3794816Z Uninstalling psutil-5.9.8: 2025-09-07T09:49:44.3800139Z Successfully uninstalled psutil-5.9.8 2025-09-07T09:49:44.5063401Z Successfully installed boto3-1.35.33 botocore-1.35.99 nvidia-ml-py-12.575.51 psutil-7.0.0 pynvml-12.0.0 s3transfer-0.10.4 2025-09-07T09:49:44.5966572Z + DEVICE_NAME= 2025-09-07T09:49:44.5966803Z + DEVICE_TYPE= 2025-09-07T09:49:44.5966983Z + command -v nvidia-smi 2025-09-07T09:49:44.5967155Z + command -v rocminfo 2025-09-07T09:49:44.5967326Z + echo DEVICE_NAME= 2025-09-07T09:49:44.5968760Z + echo DEVICE_TYPE= 2025-09-07T09:49:44.5991014Z ##[group]Run set -eux 2025-09-07T09:49:44.5991221Z set -eux 2025-09-07T09:49:44.5991393Z  2025-09-07T09:49:44.5991553Z if [[ -z "${GITHUB_TOKEN}" ]]; then 2025-09-07T09:49:44.5991782Z  echo "Missing github-token input" 2025-09-07T09:49:44.5991985Z  exit 1 2025-09-07T09:49:44.5992133Z fi 2025-09-07T09:49:44.5999209Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:49:44.5999463Z env: 2025-09-07T09:49:44.5999618Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:44.5999989Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:44.6000327Z DEVICE_NAME: 2025-09-07T09:49:44.6000477Z DEVICE_TYPE: 2025-09-07T09:49:44.6000849Z GITHUB_TOKEN: *** 2025-09-07T09:49:44.6001015Z ##[endgroup] 2025-09-07T09:49:44.6032205Z + [[ -z *** ]] 2025-09-07T09:49:44.6064927Z ##[group]Run pytorch/test-infra/.github/actions/get-workflow-job-id@main 2025-09-07T09:49:44.6065213Z with: 2025-09-07T09:49:44.6065473Z github-token: *** 2025-09-07T09:49:44.6065633Z env: 2025-09-07T09:49:44.6065795Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:44.6066101Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:44.6066416Z DEVICE_NAME: 2025-09-07T09:49:44.6066568Z DEVICE_TYPE: 2025-09-07T09:49:44.6066708Z ##[endgroup] 2025-09-07T09:49:44.6077215Z ##[group]Run set -eux 2025-09-07T09:49:44.6077412Z set -eux 2025-09-07T09:49:44.6077560Z  2025-09-07T09:49:44.6077856Z python3 "${GITHUB_ACTION_PATH}/../../scripts/get_workflow_job_id.py" "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-09-07T09:49:44.6085329Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:49:44.6085666Z env: 2025-09-07T09:49:44.6085817Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:44.6086122Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:44.6086444Z DEVICE_NAME: 2025-09-07T09:49:44.6086593Z DEVICE_TYPE: 2025-09-07T09:49:44.6086849Z GITHUB_TOKEN: *** 2025-09-07T09:49:44.6087165Z ##[endgroup] 2025-09-07T09:49:44.6114789Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/get-workflow-job-id/../../scripts/get_workflow_job_id.py 17525294857 i-0d9447236daff7b96 2025-09-07T09:49:45.0543347Z setting job-id=49775530527 2025-09-07T09:49:45.0543781Z setting job-name=inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd) 2025-09-07T09:49:45.0637565Z ##[group]Run set -eux 2025-09-07T09:49:45.0637771Z set -eux 2025-09-07T09:49:45.0637927Z  2025-09-07T09:49:45.0638078Z if [[ -n "" ]]; then 2025-09-07T09:49:45.0638276Z  source "" 2025-09-07T09:49:45.0638423Z fi 2025-09-07T09:49:45.0638562Z  2025-09-07T09:49:45.0638811Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_metadata.py" \ 2025-09-07T09:49:45.0639131Z  --schema-version "${SCHEMA_VERSION}" \ 2025-09-07T09:49:45.0639348Z  --repo "${REPO}" \ 2025-09-07T09:49:45.0639552Z  --head-branch "${HEAD_BRANCH}" \ 2025-09-07T09:49:45.0639764Z  --head-sha "${HEAD_SHA}" \ 2025-09-07T09:49:45.0639977Z  --workflow-id "${WORKFLOW_RUN_ID}" \ 2025-09-07T09:49:45.0640200Z  --run-attempt "${RUN_ATTEMPT}" \ 2025-09-07T09:49:45.0640405Z  --job-id "${JOB_ID}" \ 2025-09-07T09:49:45.0640618Z  --job-name "${JOB_NAME}" 2025-09-07T09:49:45.0647727Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:49:45.0647972Z env: 2025-09-07T09:49:45.0648123Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:45.0648427Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:45.0648748Z DEVICE_NAME: 2025-09-07T09:49:45.0648891Z DEVICE_TYPE: 2025-09-07T09:49:45.0649039Z SCHEMA_VERSION: v3 2025-09-07T09:49:45.0649211Z REPO: pytorch/pytorch 2025-09-07T09:49:45.0649385Z HEAD_BRANCH: refs/heads/main 2025-09-07T09:49:45.0649593Z HEAD_SHA: 93fb23d6fae7c4e82c4239a1033e522088742634 2025-09-07T09:49:45.0649810Z WORKFLOW_RUN_ID: 17525294857 2025-09-07T09:49:45.0649980Z RUN_ATTEMPT: 1 2025-09-07T09:49:45.0650129Z JOB_ID: 49775530527 2025-09-07T09:49:45.0650441Z JOB_NAME: inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd) 2025-09-07T09:49:45.0650786Z ##[endgroup] 2025-09-07T09:49:45.0682151Z + [[ -n '' ]] 2025-09-07T09:49:45.0683391Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_metadata.py --schema-version v3 --repo pytorch/pytorch --head-branch refs/heads/main --head-sha 93fb23d6fae7c4e82c4239a1033e522088742634 --workflow-id 17525294857 --run-attempt 1 --job-id 49775530527 --job-name 'inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd)' 2025-09-07T09:49:45.0955513Z ##[group]Run set -eux 2025-09-07T09:49:45.0955922Z set -eux 2025-09-07T09:49:45.0956177Z  2025-09-07T09:49:45.0956512Z if [[ -n "" ]]; then 2025-09-07T09:49:45.0956865Z  source "" 2025-09-07T09:49:45.0957159Z fi 2025-09-07T09:49:45.0957438Z  2025-09-07T09:49:45.0957842Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_runners_info.py" 2025-09-07T09:49:45.0965334Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:49:45.0965721Z env: 2025-09-07T09:49:45.0966047Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:45.0966479Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:45.0967130Z DEVICE_NAME: 2025-09-07T09:49:45.0967418Z DEVICE_TYPE: 2025-09-07T09:49:45.0967702Z ##[endgroup] 2025-09-07T09:49:45.0997097Z + [[ -n '' ]] 2025-09-07T09:49:45.0997851Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_runners_info.py 2025-09-07T09:49:45.1324031Z INFO:root:Fail to import torch to get the device name 2025-09-07T09:49:45.1442951Z ##[group]Run set -eux 2025-09-07T09:49:45.1443234Z set -eux 2025-09-07T09:49:45.1443571Z  2025-09-07T09:49:45.1443860Z # TODO (huydhn): Implement this part 2025-09-07T09:49:45.1444208Z echo "dependencies={}" >> "${GITHUB_OUTPUT}" 2025-09-07T09:49:45.1451759Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:49:45.1452103Z env: 2025-09-07T09:49:45.1452563Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:45.1453022Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:45.1453494Z DEVICE_NAME: 2025-09-07T09:49:45.1453807Z DEVICE_TYPE: 2025-09-07T09:49:45.1454096Z ##[endgroup] 2025-09-07T09:49:45.1483311Z + echo 'dependencies={}' 2025-09-07T09:49:45.1509615Z ##[group]Run set -eux 2025-09-07T09:49:45.1509992Z set -eux 2025-09-07T09:49:45.1510292Z  2025-09-07T09:49:45.1510643Z if [[ -n "" ]]; then 2025-09-07T09:49:45.1510958Z  source "" 2025-09-07T09:49:45.1511224Z fi 2025-09-07T09:49:45.1511552Z  2025-09-07T09:49:45.1511864Z if [[ ! -d "${BENCHMARK_RESULTS_DIR}" ]]; then 2025-09-07T09:49:45.1512262Z  echo "${BENCHMARK_RESULTS_DIR} does not exist, skipping" 2025-09-07T09:49:45.1512726Z  # We don't want the job to fail if the directory doesn't exist 2025-09-07T09:49:45.1513231Z  exit 0 2025-09-07T09:49:45.1513508Z fi 2025-09-07T09:49:45.1513851Z  2025-09-07T09:49:45.1514159Z if [[ "${DRY_RUN}" == "true" ]]; then 2025-09-07T09:49:45.1514544Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-09-07T09:49:45.1515097Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-09-07T09:49:45.1515522Z  --metadata "${BENCHMARK_METADATA}" \ 2025-09-07T09:49:45.1515905Z  --runners "${RUNNER_INFO}" \ 2025-09-07T09:49:45.1516242Z  --dependencies "${DEPENDENCIES}" \ 2025-09-07T09:49:45.1516604Z  --dry-run 2025-09-07T09:49:45.1516835Z else 2025-09-07T09:49:45.1517066Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-09-07T09:49:45.1517404Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-09-07T09:49:45.1517665Z  --metadata "${BENCHMARK_METADATA}" \ 2025-09-07T09:49:45.1517897Z  --runners "${RUNNER_INFO}" \ 2025-09-07T09:49:45.1518112Z  --dependencies "${DEPENDENCIES}" 2025-09-07T09:49:45.1518314Z fi 2025-09-07T09:49:45.1525255Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:49:45.1525494Z env: 2025-09-07T09:49:45.1525639Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:45.1525944Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:45.1526255Z DEVICE_NAME: 2025-09-07T09:49:45.1526413Z DEVICE_TYPE: 2025-09-07T09:49:45.1526578Z BENCHMARK_RESULTS_DIR: test/test-reports 2025-09-07T09:49:45.1526771Z DRY_RUN: false 2025-09-07T09:49:45.1527581Z BENCHMARK_METADATA: {"timestamp": 1757238585, "schema_version": "v3", "name": "inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd)", "repo": "pytorch/pytorch", "head_branch": "refs/heads/main", "head_sha": "93fb23d6fae7c4e82c4239a1033e522088742634", "workflow_id": 17525294857, "run_attempt": 1, "job_id": 49775530527} 2025-09-07T09:49:45.1528622Z RUNNER_INFO: [{"cpu_info": "x86_64", "cpu_count": 96, "avail_mem_in_gb": 369, "extra_info": {"hostname": "ip-10-0-57-198.ec2.internal"}, "name": "", "type": ""}] 2025-09-07T09:49:45.1529113Z DEPENDENCIES: {} 2025-09-07T09:49:45.1529275Z ##[endgroup] 2025-09-07T09:49:45.1555931Z + [[ -n '' ]] 2025-09-07T09:49:45.1556118Z + [[ ! -d test/test-reports ]] 2025-09-07T09:49:45.1556302Z + [[ false == \t\r\u\e ]] 2025-09-07T09:49:45.1558176Z + python3 /home/ec2-user/actions-runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/upload_benchmark_results.py --benchmark-results-dir test/test-reports --metadata '{"timestamp": 1757238585, "schema_version": "v3", "name": "inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd)", "repo": "pytorch/pytorch", "head_branch": "refs/heads/main", "head_sha": "93fb23d6fae7c4e82c4239a1033e522088742634", "workflow_id": 17525294857, "run_attempt": 1, "job_id": 49775530527}' --runners '[{"cpu_info": "x86_64", "cpu_count": 96, "avail_mem_in_gb": 369, "extra_info": {"hostname": "ip-10-0-57-198.ec2.internal"}, "name": "", "type": ""}]' --dependencies '{}' 2025-09-07T09:49:45.2668415Z INFO:root:Upload test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525294857/49775530527/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.json 2025-09-07T09:49:45.2950254Z INFO:botocore.credentials:Found credentials from IAM Role: gh-ci-github-action-runners-runner-role 2025-09-07T09:49:45.4490138Z INFO:root:Upload test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525294857/49775530527/inductor_dynamic_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.json 2025-09-07T09:49:45.5256957Z INFO:root:Upload test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525294857/49775530527/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.json 2025-09-07T09:49:45.6117090Z INFO:root:Upload test/test-reports/inductor_export_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525294857/49775530527/inductor_export_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.json 2025-09-07T09:49:45.7111758Z INFO:root:Upload test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525294857/49775530527/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.json 2025-09-07T09:49:45.7896491Z INFO:root:Upload test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_x86_zen_performance.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525294857/49775530527/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_x86_zen_performance.json 2025-09-07T09:49:45.9113607Z INFO:root:Upload test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_x86_zen_performance_compilation_metrics.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525294857/49775530527/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_x86_zen_performance_compilation_metrics.json 2025-09-07T09:49:46.0193871Z INFO:root:Upload test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_x86_zen_performance.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525294857/49775530527/inductor_dynamic_torchbench_bfloat16_inference_cpu_x86_zen_performance.json 2025-09-07T09:49:46.1210589Z INFO:root:Upload test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_x86_zen_performance_compilation_metrics.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525294857/49775530527/inductor_dynamic_torchbench_bfloat16_inference_cpu_x86_zen_performance_compilation_metrics.json 2025-09-07T09:49:46.2179284Z INFO:root:Upload test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_x86_zen_performance.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525294857/49775530527/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_x86_zen_performance.json 2025-09-07T09:49:46.3531781Z INFO:root:Upload test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_x86_zen_performance_compilation_metrics.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525294857/49775530527/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_x86_zen_performance_compilation_metrics.json 2025-09-07T09:49:46.4908269Z INFO:root:Upload test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_x86_zen_performance.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525294857/49775530527/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_x86_zen_performance.json 2025-09-07T09:49:46.6270742Z INFO:root:Upload test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_x86_zen_performance_compilation_metrics.json to s3://ossci-benchmarks/v3/pytorch/pytorch/17525294857/49775530527/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_x86_zen_performance_compilation_metrics.json 2025-09-07T09:49:46.7607100Z ##[group]Run cat test/**/*_toprint.log || true 2025-09-07T09:49:46.7607376Z cat test/**/*_toprint.log || true 2025-09-07T09:49:46.7614380Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:49:46.7614617Z env: 2025-09-07T09:49:46.7614766Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:46.7615069Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:46.7615378Z DEVICE_NAME: 2025-09-07T09:49:46.7615526Z DEVICE_TYPE: 2025-09-07T09:49:46.7615671Z ##[endgroup] 2025-09-07T09:49:46.7699287Z cat: 'test/**/*_toprint.log': No such file or directory 2025-09-07T09:49:46.7723842Z ##[group]Run kill "$MONITOR_SCRIPT_PID" 2025-09-07T09:49:46.7724116Z kill "$MONITOR_SCRIPT_PID" 2025-09-07T09:49:46.7731160Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:49:46.7731396Z env: 2025-09-07T09:49:46.7731545Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:46.7731861Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:46.7732171Z DEVICE_NAME: 2025-09-07T09:49:46.7732323Z DEVICE_TYPE: 2025-09-07T09:49:46.7732474Z MONITOR_SCRIPT_PID: 57444 2025-09-07T09:49:46.7732667Z ##[endgroup] 2025-09-07T09:49:46.7834628Z Prepare all required actions 2025-09-07T09:49:46.7834956Z Getting action download info 2025-09-07T09:49:46.9287482Z Download action repository 'seemethere/upload-artifact-s3@v5' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-09-07T09:49:47.1899403Z Download action repository 'actions/upload-artifact@v4' (SHA:ea165f8d65b6e75b540449e92b4886f43607fa02) 2025-09-07T09:49:47.5909181Z ##[group]Run ./.github/actions/upload-test-artifacts 2025-09-07T09:49:47.5909428Z with: 2025-09-07T09:49:47.5909705Z file-suffix: test-inductor_torchbench_perf_cpu_x86_zen-2-4-linux.24xlarge.amd_49775530527 2025-09-07T09:49:47.5910030Z s3-bucket: gha-artifacts 2025-09-07T09:49:47.5910195Z env: 2025-09-07T09:49:47.5910347Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:47.5910644Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:47.5910963Z DEVICE_NAME: 2025-09-07T09:49:47.5911112Z DEVICE_TYPE: 2025-09-07T09:49:47.5911252Z ##[endgroup] 2025-09-07T09:49:47.5927757Z ##[group]Run # Remove any previous test jsons if they exist 2025-09-07T09:49:47.5928047Z # Remove any previous test jsons if they exist 2025-09-07T09:49:47.5928278Z rm -f test-jsons-*.zip 2025-09-07T09:49:47.5928548Z zip -r "test-jsons-${FILE_SUFFIX}.zip" test/test-reports -i '*.json' 2025-09-07T09:49:47.5935466Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:49:47.5935704Z env: 2025-09-07T09:49:47.5935846Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:47.5936232Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:47.5936544Z DEVICE_NAME: 2025-09-07T09:49:47.5936695Z DEVICE_TYPE: 2025-09-07T09:49:47.5937067Z FILE_SUFFIX: test-inductor_torchbench_perf_cpu_x86_zen-2-4-linux.24xlarge.amd_49775530527 2025-09-07T09:49:47.5937363Z ##[endgroup] 2025-09-07T09:49:47.6075397Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.json (deflated 99%) 2025-09-07T09:49:47.6102725Z adding: test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.json (deflated 99%) 2025-09-07T09:49:47.6130467Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.json (deflated 99%) 2025-09-07T09:49:47.6157928Z adding: test/test-reports/inductor_export_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.json (deflated 99%) 2025-09-07T09:49:47.6178494Z adding: test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.json (deflated 99%) 2025-09-07T09:49:47.6220535Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_x86_zen_performance.json (deflated 99%) 2025-09-07T09:49:47.6315276Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_x86_zen_performance_compilation_metrics.json (deflated 99%) 2025-09-07T09:49:47.6357052Z adding: test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_x86_zen_performance.json (deflated 99%) 2025-09-07T09:49:47.6452444Z adding: test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_x86_zen_performance_compilation_metrics.json (deflated 99%) 2025-09-07T09:49:47.6493934Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_x86_zen_performance.json (deflated 99%) 2025-09-07T09:49:47.6588783Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_x86_zen_performance_compilation_metrics.json (deflated 99%) 2025-09-07T09:49:47.6619292Z adding: test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_x86_zen_performance.json (deflated 99%) 2025-09-07T09:49:47.6664403Z adding: test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_x86_zen_performance_compilation_metrics.json (deflated 99%) 2025-09-07T09:49:47.6689566Z ##[group]Run # Remove any previous test reports if they exist 2025-09-07T09:49:47.6690108Z # Remove any previous test reports if they exist 2025-09-07T09:49:47.6690367Z rm -f test-reports-*.zip 2025-09-07T09:49:47.6690658Z zip -r "test-reports-${FILE_SUFFIX}.zip" test/test-reports -i '*.xml' -i '*.csv' 2025-09-07T09:49:47.6697918Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:49:47.6698160Z env: 2025-09-07T09:49:47.6698309Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:47.6698629Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:47.6698953Z DEVICE_NAME: 2025-09-07T09:49:47.6699099Z DEVICE_TYPE: 2025-09-07T09:49:47.6699368Z FILE_SUFFIX: test-inductor_torchbench_perf_cpu_x86_zen-2-4-linux.24xlarge.amd_49775530527 2025-09-07T09:49:47.6699681Z ##[endgroup] 2025-09-07T09:49:47.6763871Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.csv (deflated 59%) 2025-09-07T09:49:47.6764472Z adding: test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.csv (deflated 58%) 2025-09-07T09:49:47.6765029Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.csv (deflated 59%) 2025-09-07T09:49:47.6765567Z adding: test/test-reports/inductor_export_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.csv (deflated 63%) 2025-09-07T09:49:47.6766107Z adding: test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_x86_zen_accuracy.csv (deflated 71%) 2025-09-07T09:49:47.6766682Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_x86_zen_performance.csv (deflated 50%) 2025-09-07T09:49:47.6768911Z adding: test/test-reports/inductor_no_cudagraphs_torchbench_bfloat16_inference_cpu_x86_zen_performance_compilation_metrics.csv (deflated 51%) 2025-09-07T09:49:47.6769647Z adding: test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_x86_zen_performance.csv (deflated 50%) 2025-09-07T09:49:47.6772149Z adding: test/test-reports/inductor_dynamic_torchbench_bfloat16_inference_cpu_x86_zen_performance_compilation_metrics.csv (deflated 51%) 2025-09-07T09:49:47.6772763Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_x86_zen_performance.csv (deflated 50%) 2025-09-07T09:49:47.6820415Z adding: test/test-reports/inductor_cpp_wrapper_torchbench_bfloat16_inference_cpu_x86_zen_performance_compilation_metrics.csv (deflated 51%) 2025-09-07T09:49:47.6821271Z adding: test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_x86_zen_performance.csv (deflated 52%) 2025-09-07T09:49:47.6821896Z adding: test/test-reports/inductor_aot_inductor_torchbench_bfloat16_inference_cpu_x86_zen_performance_compilation_metrics.csv (deflated 49%) 2025-09-07T09:49:47.6836994Z ##[group]Run # Remove any previous usage logs if they exist 2025-09-07T09:49:47.6837288Z # Remove any previous usage logs if they exist 2025-09-07T09:49:47.6837521Z rm -f logs-*.zip 2025-09-07T09:49:47.6837752Z zip "logs-${FILE_SUFFIX}.zip" 'usage_log.txt' || true 2025-09-07T09:49:47.6838070Z zip -r "logs-${FILE_SUFFIX}.zip" test/test-reports -i '*.log' || true 2025-09-07T09:49:47.6844208Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:49:47.6844453Z env: 2025-09-07T09:49:47.6844605Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:47.6844912Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:47.6845226Z DEVICE_NAME: 2025-09-07T09:49:47.6845381Z DEVICE_TYPE: 2025-09-07T09:49:47.6845648Z FILE_SUFFIX: test-inductor_torchbench_perf_cpu_x86_zen-2-4-linux.24xlarge.amd_49775530527 2025-09-07T09:49:47.6845954Z ##[endgroup] 2025-09-07T09:49:47.6942505Z adding: usage_log.txt (deflated 96%) 2025-09-07T09:49:47.6956599Z 2025-09-07T09:49:47.6956879Z zip error: Nothing to do! (logs-test-inductor_torchbench_perf_cpu_x86_zen-2-4-linux.24xlarge.amd_49775530527.zip) 2025-09-07T09:49:47.6977804Z ##[group]Run # Remove any previous debugging artifacts if they exist 2025-09-07T09:49:47.6978187Z # Remove any previous debugging artifacts if they exist 2025-09-07T09:49:47.6978455Z rm -f debug-*.zip 2025-09-07T09:49:47.6978643Z if [ -d 'test/debug' ]; then 2025-09-07T09:49:47.6978870Z  zip -r "debug-${FILE_SUFFIX}.zip" test/debug 2025-09-07T09:49:47.6979083Z fi 2025-09-07T09:49:47.6986049Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:49:47.6986300Z env: 2025-09-07T09:49:47.6986447Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:47.6986759Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:47.6987088Z DEVICE_NAME: 2025-09-07T09:49:47.6987249Z DEVICE_TYPE: 2025-09-07T09:49:47.6987513Z FILE_SUFFIX: test-inductor_torchbench_perf_cpu_x86_zen-2-4-linux.24xlarge.amd_49775530527 2025-09-07T09:49:47.6987814Z ##[endgroup] 2025-09-07T09:49:47.7059082Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-09-07T09:49:47.7059315Z with: 2025-09-07T09:49:47.7059476Z s3-bucket: gha-artifacts 2025-09-07T09:49:47.7059699Z s3-prefix: pytorch/pytorch/17525294857/1/artifact 2025-09-07T09:49:47.7059985Z retention-days: 14 2025-09-07T09:49:47.7060149Z if-no-files-found: warn 2025-09-07T09:49:47.7060344Z path: test-jsons-*.zip 2025-09-07T09:49:47.7060523Z name: artifact 2025-09-07T09:49:47.7060683Z region: us-east-1 2025-09-07T09:49:47.7060832Z env: 2025-09-07T09:49:47.7061007Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:47.7061318Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:47.7061635Z DEVICE_NAME: 2025-09-07T09:49:47.7061781Z DEVICE_TYPE: 2025-09-07T09:49:47.7061933Z ##[endgroup] 2025-09-07T09:49:47.9847141Z NOTE: s3-prefix specified, ignoring name parameter 2025-09-07T09:49:47.9847474Z With the provided path, there will be 1 file uploaded 2025-09-07T09:49:47.9847768Z Uploading to s3 prefix: pytorch/pytorch/17525294857/1/artifact 2025-09-07T09:49:47.9905372Z Starting upload of test-jsons-test-inductor_torchbench_perf_cpu_x86_zen-2-4-linux.24xlarge.amd_49775530527.zip 2025-09-07T09:49:48.1117822Z Finished upload of test-jsons-test-inductor_torchbench_perf_cpu_x86_zen-2-4-linux.24xlarge.amd_49775530527.zip 2025-09-07T09:49:48.1329809Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-09-07T09:49:48.1330040Z with: 2025-09-07T09:49:48.1330195Z s3-bucket: gha-artifacts 2025-09-07T09:49:48.1330536Z s3-prefix: pytorch/pytorch/17525294857/1/artifact 2025-09-07T09:49:48.1330764Z retention-days: 14 2025-09-07T09:49:48.1330923Z if-no-files-found: error 2025-09-07T09:49:48.1331098Z path: test-reports-*.zip 2025-09-07T09:49:48.1331262Z name: artifact 2025-09-07T09:49:48.1331411Z region: us-east-1 2025-09-07T09:49:48.1331563Z env: 2025-09-07T09:49:48.1331706Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:48.1332009Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:48.1332334Z DEVICE_NAME: 2025-09-07T09:49:48.1332487Z DEVICE_TYPE: 2025-09-07T09:49:48.1332631Z ##[endgroup] 2025-09-07T09:49:48.4007370Z NOTE: s3-prefix specified, ignoring name parameter 2025-09-07T09:49:48.4007696Z With the provided path, there will be 1 file uploaded 2025-09-07T09:49:48.4007990Z Uploading to s3 prefix: pytorch/pytorch/17525294857/1/artifact 2025-09-07T09:49:48.4064788Z Starting upload of test-reports-test-inductor_torchbench_perf_cpu_x86_zen-2-4-linux.24xlarge.amd_49775530527.zip 2025-09-07T09:49:48.5144202Z Finished upload of test-reports-test-inductor_torchbench_perf_cpu_x86_zen-2-4-linux.24xlarge.amd_49775530527.zip 2025-09-07T09:49:48.5355807Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-09-07T09:49:48.5356030Z with: 2025-09-07T09:49:48.5356198Z s3-bucket: gha-artifacts 2025-09-07T09:49:48.5356420Z s3-prefix: pytorch/pytorch/17525294857/1/artifact 2025-09-07T09:49:48.5356643Z retention-days: 14 2025-09-07T09:49:48.5356802Z if-no-files-found: ignore 2025-09-07T09:49:48.5356978Z path: logs-*.zip 2025-09-07T09:49:48.5357334Z name: artifact 2025-09-07T09:49:48.5357489Z region: us-east-1 2025-09-07T09:49:48.5357632Z env: 2025-09-07T09:49:48.5357774Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:48.5358083Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:48.5358406Z DEVICE_NAME: 2025-09-07T09:49:48.5358553Z DEVICE_TYPE: 2025-09-07T09:49:48.5358702Z ##[endgroup] 2025-09-07T09:49:48.8055113Z NOTE: s3-prefix specified, ignoring name parameter 2025-09-07T09:49:48.8055429Z With the provided path, there will be 1 file uploaded 2025-09-07T09:49:48.8055713Z Uploading to s3 prefix: pytorch/pytorch/17525294857/1/artifact 2025-09-07T09:49:48.8112663Z Starting upload of logs-test-inductor_torchbench_perf_cpu_x86_zen-2-4-linux.24xlarge.amd_49775530527.zip 2025-09-07T09:49:48.9207493Z Finished upload of logs-test-inductor_torchbench_perf_cpu_x86_zen-2-4-linux.24xlarge.amd_49775530527.zip 2025-09-07T09:49:48.9420368Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-09-07T09:49:48.9420612Z with: 2025-09-07T09:49:48.9420767Z s3-bucket: gha-artifacts 2025-09-07T09:49:48.9420981Z s3-prefix: pytorch/pytorch/17525294857/1/artifact 2025-09-07T09:49:48.9421203Z retention-days: 14 2025-09-07T09:49:48.9421368Z if-no-files-found: ignore 2025-09-07T09:49:48.9421546Z path: debug-*.zip 2025-09-07T09:49:48.9421688Z name: artifact 2025-09-07T09:49:48.9421838Z region: us-east-1 2025-09-07T09:49:48.9421995Z env: 2025-09-07T09:49:48.9422134Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:48.9422430Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:48.9422752Z DEVICE_NAME: 2025-09-07T09:49:48.9422906Z DEVICE_TYPE: 2025-09-07T09:49:48.9423171Z ##[endgroup] 2025-09-07T09:49:49.2060229Z No files were found with the provided path: debug-*.zip. No artifacts will be uploaded. 2025-09-07T09:49:49.2326056Z ##[group]Run # shellcheck disable=SC2156 2025-09-07T09:49:49.2326330Z # shellcheck disable=SC2156 2025-09-07T09:49:49.2326700Z find . -iname "core.[1-9]*" -exec docker exec "${DOCKER_CONTAINER_ID}" sh -c "gdb python {} -ex 'bt' -ex 'q'" \; 2025-09-07T09:49:49.2333961Z shell: /usr/bin/bash -e {0} 2025-09-07T09:49:49.2334146Z env: 2025-09-07T09:49:49.2334287Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:49.2334597Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:49.2334914Z DEVICE_NAME: 2025-09-07T09:49:49.2335166Z DEVICE_TYPE: 2025-09-07T09:49:49.2335308Z ##[endgroup] 2025-09-07T09:49:49.5452068Z Prepare all required actions 2025-09-07T09:49:49.5452339Z Getting action download info 2025-09-07T09:49:49.6849870Z ##[group]Run ./.github/actions/upload-utilization-stats 2025-09-07T09:49:49.6850115Z with: 2025-09-07T09:49:49.6850263Z job_id: 49775530527 2025-09-07T09:49:49.6850578Z job_name: inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd) 2025-09-07T09:49:49.6850948Z workflow_name: inductor-perf-nightly-x86-zen 2025-09-07T09:49:49.6851167Z workflow_run_id: 17525294857 2025-09-07T09:49:49.6851339Z workflow_attempt: 1 2025-09-07T09:49:49.6851492Z env: 2025-09-07T09:49:49.6851635Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:49.6851933Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:49.6852247Z DEVICE_NAME: 2025-09-07T09:49:49.6852397Z DEVICE_TYPE: 2025-09-07T09:49:49.6852546Z ##[endgroup] 2025-09-07T09:49:49.6864369Z ##[group]Run echo "workflow_id: 17525294857" 2025-09-07T09:49:49.6864630Z echo "workflow_id: 17525294857" 2025-09-07T09:49:49.6864846Z echo "workflow_attempt: 1" 2025-09-07T09:49:49.6865091Z echo "workflow_Name: inductor-perf-nightly-x86-zen" 2025-09-07T09:49:49.6865333Z echo "job_id: 49775530527" 2025-09-07T09:49:49.6865694Z echo "job_name: inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd)" 2025-09-07T09:49:49.6866294Z echo "artifact_prefix: " 2025-09-07T09:49:49.6866497Z python3 --version 2025-09-07T09:49:49.6873675Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:49:49.6873912Z env: 2025-09-07T09:49:49.6874060Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:49.6874361Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:49.6874680Z DEVICE_NAME: 2025-09-07T09:49:49.6874831Z DEVICE_TYPE: 2025-09-07T09:49:49.6874971Z ##[endgroup] 2025-09-07T09:49:49.6900821Z workflow_id: 17525294857 2025-09-07T09:49:49.6901001Z workflow_attempt: 1 2025-09-07T09:49:49.6901205Z workflow_Name: inductor-perf-nightly-x86-zen 2025-09-07T09:49:49.6901424Z job_id: 49775530527 2025-09-07T09:49:49.6901741Z job_name: inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd) 2025-09-07T09:49:49.6902082Z artifact_prefix: 2025-09-07T09:49:49.6916622Z Python 3.9.23 2025-09-07T09:49:49.6947864Z ##[group]Run nick-fields/retry@v3.0.0 2025-09-07T09:49:49.6948082Z with: 2025-09-07T09:49:49.6948217Z shell: bash 2025-09-07T09:49:49.6948368Z timeout_minutes: 5 2025-09-07T09:49:49.6948530Z max_attempts: 5 2025-09-07T09:49:49.6948698Z retry_wait_seconds: 30 2025-09-07T09:49:49.6949042Z command: set -eu python3 -m pip install python-dateutil==2.8.2 boto3==1.35.42 pandas==2.1.3 dataclasses_json==0.6.7 2025-09-07T09:49:49.6949408Z polling_interval_seconds: 1 2025-09-07T09:49:49.6949612Z warning_on_retry: true 2025-09-07T09:49:49.6949805Z continue_on_error: false 2025-09-07T09:49:49.6949973Z env: 2025-09-07T09:49:49.6950109Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:49.6950408Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:49.6950856Z DEVICE_NAME: 2025-09-07T09:49:49.6951007Z DEVICE_TYPE: 2025-09-07T09:49:49.6951143Z ##[endgroup] 2025-09-07T09:49:49.9545213Z Defaulting to user installation because normal site-packages is not writeable 2025-09-07T09:49:50.0154190Z Collecting python-dateutil==2.8.2 2025-09-07T09:49:50.0454453Z Downloading python_dateutil-2.8.2-py2.py3-none-any.whl (247 kB) 2025-09-07T09:49:50.7129406Z Collecting boto3==1.35.42 2025-09-07T09:49:50.7220061Z Downloading boto3-1.35.42-py3-none-any.whl (139 kB) 2025-09-07T09:49:51.0747637Z Collecting pandas==2.1.3 2025-09-07T09:49:51.0844527Z Downloading pandas-2.1.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (12.3 MB) 2025-09-07T09:49:51.1927257Z Requirement already satisfied: dataclasses_json==0.6.7 in /home/ec2-user/.local/lib/python3.9/site-packages (0.6.7) 2025-09-07T09:49:51.1937947Z Requirement already satisfied: six>=1.5 in /usr/lib/python3.9/site-packages (from python-dateutil==2.8.2) (1.15.0) 2025-09-07T09:49:51.1968704Z Requirement already satisfied: botocore<1.36.0,>=1.35.42 in /home/ec2-user/.local/lib/python3.9/site-packages (from boto3==1.35.42) (1.35.99) 2025-09-07T09:49:51.1972336Z Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /usr/lib/python3.9/site-packages (from boto3==1.35.42) (0.10.0) 2025-09-07T09:49:51.1974837Z Requirement already satisfied: s3transfer<0.11.0,>=0.10.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from boto3==1.35.42) (0.10.4) 2025-09-07T09:49:51.2610699Z Collecting tzdata>=2022.1 2025-09-07T09:49:51.2698121Z Downloading tzdata-2025.2-py2.py3-none-any.whl (347 kB) 2025-09-07T09:49:51.8284387Z Collecting numpy<2,>=1.22.4 2025-09-07T09:49:51.8376523Z Downloading numpy-1.26.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (18.2 MB) 2025-09-07T09:49:51.9780934Z Requirement already satisfied: pytz>=2020.1 in /usr/lib/python3.9/site-packages (from pandas==2.1.3) (2022.7.1) 2025-09-07T09:49:51.9801818Z Requirement already satisfied: marshmallow<4.0.0,>=3.18.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from dataclasses_json==0.6.7) (3.26.1) 2025-09-07T09:49:51.9804830Z Requirement already satisfied: typing-inspect<1,>=0.4.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from dataclasses_json==0.6.7) (0.9.0) 2025-09-07T09:49:51.9847245Z Requirement already satisfied: urllib3<1.27,>=1.25.4 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.42->boto3==1.35.42) (1.25.10) 2025-09-07T09:49:51.9923636Z Requirement already satisfied: packaging>=17.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from marshmallow<4.0.0,>=3.18.0->dataclasses_json==0.6.7) (25.0) 2025-09-07T09:49:51.9993504Z Requirement already satisfied: mypy-extensions>=0.3.0 in /home/ec2-user/.local/lib/python3.9/site-packages (from typing-inspect<1,>=0.4.0->dataclasses_json==0.6.7) (1.1.0) 2025-09-07T09:49:51.9995850Z Requirement already satisfied: typing-extensions>=3.7.4 in /home/ec2-user/.local/lib/python3.9/site-packages (from typing-inspect<1,>=0.4.0->dataclasses_json==0.6.7) (4.15.0) 2025-09-07T09:49:52.1362304Z Installing collected packages: python-dateutil, tzdata, numpy, pandas, boto3 2025-09-07T09:49:55.8870493Z Attempting uninstall: boto3 2025-09-07T09:49:55.8870822Z Found existing installation: boto3 1.35.33 2025-09-07T09:49:55.8961321Z Uninstalling boto3-1.35.33: 2025-09-07T09:49:55.8973446Z Successfully uninstalled boto3-1.35.33 2025-09-07T09:49:55.9419179Z Successfully installed boto3-1.35.42 numpy-1.26.4 pandas-2.1.3 python-dateutil-2.8.2 tzdata-2025.2 2025-09-07T09:49:56.7622889Z Command completed after 1 attempt(s). 2025-09-07T09:49:56.7679873Z ##[group]Run python3 -m tools.stats.upload_utilization_stats.upload_utilization_stats \ 2025-09-07T09:49:56.7680342Z python3 -m tools.stats.upload_utilization_stats.upload_utilization_stats \ 2025-09-07T09:49:56.7680651Z  --workflow-run-id "17525294857" \ 2025-09-07T09:49:56.7680919Z  --workflow-name "inductor-perf-nightly-x86-zen" \ 2025-09-07T09:49:56.7681289Z  --workflow-run-attempt "1" \ 2025-09-07T09:49:56.7681497Z  --job-id "49775530527" \ 2025-09-07T09:49:56.7681861Z  --job-name "inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd)" \ 2025-09-07T09:49:56.7682219Z  --local-path "" \ 2025-09-07T09:49:56.7682413Z  --artifact-prefix "" 2025-09-07T09:49:56.7690454Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:49:56.7690692Z env: 2025-09-07T09:49:56.7690837Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:56.7691139Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:56.7691449Z DEVICE_NAME: 2025-09-07T09:49:56.7691691Z DEVICE_TYPE: 2025-09-07T09:49:56.7691834Z ##[endgroup] 2025-09-07T09:49:57.6565015Z repo: pytorch/pytorch 2025-09-07T09:49:57.6565355Z Search for test log in s3 bucket: ossci-utilization 2025-09-07T09:49:57.6565747Z Downloading logs-test-inductor_torchbench_perf_cpu_x86_zen-2-4-linux.24xlarge.amd_49775530527.zip 2025-09-07T09:49:57.6566285Z extracting usage_log.txt from zip file logs-test-inductor_torchbench_perf_cpu_x86_zen-2-4-linux.24xlarge.amd_49775530527.zip 2025-09-07T09:49:57.6566683Z Converted Log Model: UtilizationMetadata: 2025-09-07T09:49:57.6567602Z UtilizationMetadata(level='metadata', workflow_id='17525294857', job_id='49775530527', workflow_name='inductor-perf-nightly-x86-zen', job_name='inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd)', usage_collect_interval=4.0, data_model_version=1.5, start_at=1757231017, gpu_count=0, cpu_count=96, gpu_type=None, error=None) 2025-09-07T09:49:57.6568569Z [Db Segments] detected pytest cmd: 14, generated segments: 14 2025-09-07T09:49:57.6568832Z [db model] Peek db timeseries 2025-09-07T09:49:57.6569009Z :{ 2025-09-07T09:49:57.6569149Z "created_at": 1757238597, 2025-09-07T09:49:57.6569328Z "type": "utilization", 2025-09-07T09:49:57.6569486Z "tags": [ 2025-09-07T09:49:57.6569627Z "record" 2025-09-07T09:49:57.6569766Z ], 2025-09-07T09:49:57.6569906Z "time_stamp": 1757231017, 2025-09-07T09:49:57.6570078Z "repo": "pytorch/pytorch", 2025-09-07T09:49:57.6570265Z "workflow_id": 17525294857, 2025-09-07T09:49:57.6570443Z "run_attempt": 1, 2025-09-07T09:49:57.6570604Z "job_id": 49775530527, 2025-09-07T09:49:57.6570808Z "workflow_name": "inductor-perf-nightly-x86-zen", 2025-09-07T09:49:57.6571188Z "job_name": "inductor-test-nightly / test (inductor_torchbench_perf_cpu_x86_zen, 2, 4, linux.24xlarge.amd)", 2025-09-07T09:49:57.6571527Z "json_data": "{}" 2025-09-07T09:49:57.6571676Z } 2025-09-07T09:49:57.6571984Z Writing 1 documents to S3 ossci-utilization/util_metadata/v_1.5/pytorch/pytorch/17525294857/1/49775530527/metadata 2025-09-07T09:49:57.6572548Z Done! Finish writing document to S3 ossci-utilization/util_metadata/v_1.5/pytorch/pytorch/17525294857/1/49775530527/metadata 2025-09-07T09:49:57.6573113Z Writing 498 documents to S3 ossci-utilization/util_timeseries/v_1.5/pytorch/pytorch/17525294857/1/49775530527/time_series 2025-09-07T09:49:57.6573678Z Done! Finish writing document to S3 ossci-utilization/util_timeseries/v_1.5/pytorch/pytorch/17525294857/1/49775530527/time_series 2025-09-07T09:49:57.7445370Z ##[group]Run pytorch/test-infra/.github/actions/teardown-linux@main 2025-09-07T09:49:57.7445689Z with: 2025-09-07T09:49:57.7446108Z env: 2025-09-07T09:49:57.7446254Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:57.7446568Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:57.7446895Z DEVICE_NAME: 2025-09-07T09:49:57.7447049Z DEVICE_TYPE: 2025-09-07T09:49:57.7447186Z ##[endgroup] 2025-09-07T09:49:57.7468758Z ##[group]Run set -eou pipefail 2025-09-07T09:49:57.7469044Z set -eou pipefail 2025-09-07T09:49:57.7469226Z  2025-09-07T09:49:57.7469462Z echo "Holding runner for 2 hours until all ssh sessions have logged out" 2025-09-07T09:49:57.7469863Z for _ in $(seq 1440); do 2025-09-07T09:49:57.7470078Z  # Break if no ssh session exists anymore 2025-09-07T09:49:57.7470303Z  if [ "$(who)" = "" ]; then 2025-09-07T09:49:57.7470491Z  break 2025-09-07T09:49:57.7470674Z  fi 2025-09-07T09:49:57.7470828Z  echo "." 2025-09-07T09:49:57.7470983Z  sleep 5 2025-09-07T09:49:57.7471137Z done 2025-09-07T09:49:57.7478409Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:49:57.7478647Z env: 2025-09-07T09:49:57.7478786Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:57.7479093Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:57.7479517Z DEVICE_NAME: 2025-09-07T09:49:57.7479665Z DEVICE_TYPE: 2025-09-07T09:49:57.7479804Z ##[endgroup] 2025-09-07T09:49:57.7506260Z Holding runner for 2 hours until all ssh sessions have logged out 2025-09-07T09:49:57.7647222Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-09-07T09:49:57.7647619Z # ignore expansion of "docker ps -q" since it could be empty 2025-09-07T09:49:57.7647902Z # shellcheck disable=SC2046 2025-09-07T09:49:57.7648147Z docker stop $(docker ps -q) || true 2025-09-07T09:49:57.7648364Z # Prune all of the docker images 2025-09-07T09:49:57.7648576Z docker system prune -af 2025-09-07T09:49:57.7656272Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:49:57.7656521Z env: 2025-09-07T09:49:57.7656662Z GIT_DEFAULT_BRANCH: main 2025-09-07T09:49:57.7656984Z DOCKER_CONTAINER_ID: de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:49:57.7657324Z DEVICE_NAME: 2025-09-07T09:49:57.7657549Z DEVICE_TYPE: 2025-09-07T09:49:57.7657696Z ##[endgroup] 2025-09-07T09:50:09.0281794Z de0caba1838c 2025-09-07T09:50:10.5891194Z Deleted Containers: 2025-09-07T09:50:10.5891532Z de0caba1838c50b38cd32bd07459a7acfcd20c5f8c2aeab4a1af2a5c7c9a3fac 2025-09-07T09:50:10.5891774Z 2025-09-07T09:50:26.4665019Z Deleted Images: 2025-09-07T09:50:26.4665678Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-py3-gcc11-inductor-benchmarks-ae53c6842aa4c2407d0ad976491ca941c2635c77 2025-09-07T09:50:26.4666493Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image@sha256:383efb45082f20b8c808cb0ba4df693a01359592233f641f1f486911ac320a9a 2025-09-07T09:50:26.4667056Z deleted: sha256:662d8c9dfc7db2f5d004293de4f2b7647941dee4c916479ef082d17fcdfd9c47 2025-09-07T09:50:26.4667455Z deleted: sha256:ea5ad443c754124b3a5a209c2663376b4c156947edef1b982a336148bbf9114d 2025-09-07T09:50:26.4667826Z deleted: sha256:284be7504f072e0c04da4e2190e8d0e1de73835ed67be81f3ddd7eafd5d06a3a 2025-09-07T09:50:26.4668224Z deleted: sha256:2f49ff4be65f7ca55de8d7028fb3df7d08232a9f043aa7ba27d9393724286281 2025-09-07T09:50:26.4668618Z deleted: sha256:f63b503fdd1cca198aecefb9eef7ffbeb5fbc723f2a8462f50316e56cd403cbc 2025-09-07T09:50:26.4669025Z deleted: sha256:f9d46e08457013f0e71d608ac3dd95b79c41120060a80baefa684048cc15574e 2025-09-07T09:50:26.4669408Z deleted: sha256:cab76e28615751b6d6a703103b1da790a67cb3a4ee2e8814de51de18ff8b595d 2025-09-07T09:50:26.4670329Z deleted: sha256:0b2d09aa482371591a32563a5db71472822abd096a347967a9bd2a177737109f 2025-09-07T09:50:26.4670719Z deleted: sha256:d306d346d5da05e9fd04284304b1637a0bf01ee97397c688d19d783d5e133de9 2025-09-07T09:50:26.4671090Z deleted: sha256:bb3381a916d410a6e304540bb0796099dc780cd11f5829e734b337e0e79acfe4 2025-09-07T09:50:26.4671455Z deleted: sha256:bcf487c27e826c092985285163fb896e3324460b1774f3eb2a66623cd31e7d87 2025-09-07T09:50:26.4671816Z deleted: sha256:7d13485a9bdc5c0e64ac5085b25f4dded75c60f74090369c1b6f3f546ee37e94 2025-09-07T09:50:26.4672195Z deleted: sha256:55351d98a4197542fa7c78089671f447a6ef88cc554b7fad4fc522e8d4d187b6 2025-09-07T09:50:26.4672567Z deleted: sha256:f884bc0c4f9a994f3b3f1d82205f3a7014b05c84ad0c1c2fa3254d15a44f31e1 2025-09-07T09:50:26.4672942Z deleted: sha256:cdd16785a15239e518604ea9ea31405d5225fa6411d1c6d74d6523bcebf759ab 2025-09-07T09:50:26.4673445Z deleted: sha256:2c5bc1dc49446d7df5784578ae7c99460a93b502aa0c3b9deffbb95ec5216860 2025-09-07T09:50:26.4673819Z deleted: sha256:bae1e956be98416ce7d1a6c2c6ef0917f467238e19291786f8e1fed36fa81956 2025-09-07T09:50:26.4674192Z deleted: sha256:2cb1f002ab1126b0606999a9557b3f7f5da1e453d5376d29d95d60a979a215c4 2025-09-07T09:50:26.4674571Z deleted: sha256:25055a5f67b9bce8fac50ee1508dcb0f862ed154de5ded734e55f60edaca385f 2025-09-07T09:50:26.4674952Z deleted: sha256:98024e2dd34a5899240e41ae14f59c657cdc005040773e6ad7cfe3d67cdac7a8 2025-09-07T09:50:26.4675332Z deleted: sha256:8d2e75659096b4af8a20c3e9a6cce899b6e720f638eacdfd7d41ec8a736efdde 2025-09-07T09:50:26.4675800Z deleted: sha256:7741a6bf043548509c51c32e44734f30dfe07f91ca56c64422b004c3c0444e68 2025-09-07T09:50:26.4676180Z deleted: sha256:e2e63edbd2512e413c388888eabade05a2a7876adf20e7f0e0c3660ac3acbd3d 2025-09-07T09:50:26.4676559Z deleted: sha256:7fdea0f7711ee22084f87dc6d651598b5e5c5237de828105f698cb6a937d4c9c 2025-09-07T09:50:26.4676938Z deleted: sha256:486a2cf42f9492f291d59d48f3cec5a0a72449d8b6ad7d7a02596da237cdd154 2025-09-07T09:50:26.4677318Z deleted: sha256:a17da64c93a4939fad81a3ff6b6cb30f988176a6e0062fcf9c65e06cd9b9c3fb 2025-09-07T09:50:26.4677698Z deleted: sha256:70b4a3a917b8f95b19ae5dab6f404af8fa1c886022e4a1d785654013d5d876af 2025-09-07T09:50:26.4678078Z deleted: sha256:bd1b9d6a8aa636a67023800dcd85e4a3a7a7a21d65c6e6491d169fa65b4404a9 2025-09-07T09:50:26.4678457Z deleted: sha256:e3befcf3d3693c1d7bf0535e6e6722f0aabb0123805443ef5915dd5441ed0b00 2025-09-07T09:50:26.4678837Z deleted: sha256:4b4f846f1c4266b015f5fdf8dac5346c083c3aee2375e337172c112677c5a8c0 2025-09-07T09:50:26.4679208Z deleted: sha256:f05dc4d1350267b90e07af241a64f86a928fb3d8de75717ac04ec5a0433d042f 2025-09-07T09:50:26.4679593Z deleted: sha256:b6b4de696915fa2db09844ec9ac44dbb2940b655cd356404cf1ff03eec644dad 2025-09-07T09:50:26.4679969Z deleted: sha256:da008bbe1fc29cb35b3949040e97eb801f3264a56c4dd1b9d43a3cb54f2a39b2 2025-09-07T09:50:26.4680352Z deleted: sha256:261da5d14cad99ee11dcdaeb6055726f38fc12b7c559ee9c6d2ddc3f288f4828 2025-09-07T09:50:26.4680776Z deleted: sha256:16f900c60e70d685a85ca571ee0dada993a02217bdd6bb8b1d49169e7e28cf41 2025-09-07T09:50:26.4691201Z deleted: sha256:f57b18c5cde1d1dc553a15e1e98141d4afc0b4d0bb1182cc85b2c21bd18bb783 2025-09-07T09:50:26.4691690Z deleted: sha256:3c79105088ac60b231e4553752ee42cb6a87f9d32736b32f0c2123dddec724e7 2025-09-07T09:50:26.4692085Z deleted: sha256:df1ffff478908236efb6ceb8e05e6e078f12b864f4d24ce598cba7b961fad65c 2025-09-07T09:50:26.4692471Z deleted: sha256:8170255b562b59b76768f18a5b84b1ba887db93d3fe43b87a74bdc6be4f82014 2025-09-07T09:50:26.4692851Z deleted: sha256:c863cfe6bed704be5a54617331e27158b6f5a492dd6b9ed9c99d23db017cf5e1 2025-09-07T09:50:26.4693245Z deleted: sha256:e9e5a98c073f72c3abf9cc98724a31a3791535574ac78aeda7eb5df4580b21d0 2025-09-07T09:50:26.4693614Z deleted: sha256:0a42ac98735ca6578911218be7a7918001fe8aee1eb33d98f0d0a153d0e1102d 2025-09-07T09:50:26.4694000Z deleted: sha256:77d5a8aaa4d0fe1210dda9ac1f0fa3cf6141fea925b6240b9839d7505d021d3f 2025-09-07T09:50:26.4694388Z deleted: sha256:fa6ec46c43532dc01449df1cc403de8bb5872f859076e90658534c51c1487ef9 2025-09-07T09:50:26.4694898Z deleted: sha256:424a12dd5083283e19af48d31b7f2e33911ca8f459796f17280eaf5777a9aa25 2025-09-07T09:50:26.4695288Z deleted: sha256:8f0499601e14f1073e20ce889b45d12ab33264f9cf30359ac29dddbf58a311aa 2025-09-07T09:50:26.4695668Z deleted: sha256:5a5fae32dfb81abcd7bf374018b11e8e42a5aa39841d4b94e822d306c9af015b 2025-09-07T09:50:26.4696059Z deleted: sha256:d1bda89f22d383d38dfb7f7590b3bb202ccb91814034e7c7e2493306a10151ef 2025-09-07T09:50:26.4696442Z deleted: sha256:dbf16c1fcae146528685a8f745f9c505b24ba9ef009c42b1bd711ff7bf51b936 2025-09-07T09:50:26.4696819Z deleted: sha256:f9ec0065788f638325536a37427e2635b760a32457f20ca0acbcef6946b1041b 2025-09-07T09:50:26.4697190Z deleted: sha256:9d9911dac8fb2ff7db87329f38625d73f452dfef8822830048bbc00541c7df14 2025-09-07T09:50:26.4697638Z deleted: sha256:de4c1937129850e357b0de484d230569f628ac0bc883b12eff42932cd1e193ce 2025-09-07T09:50:26.4698067Z deleted: sha256:7b3c9e5b56a1d74226a5c1a54e5cb5e749012aa9b1d2376c6e7503757e29c35b 2025-09-07T09:50:26.4698445Z deleted: sha256:8062a6f28fc5fe2a199e1c1c40b6c43b7e29eb0c452492b47ec6900413b19cb6 2025-09-07T09:50:26.4698833Z deleted: sha256:f879aeffe6886f8da80462b571f9307aa63bb961645bec55ff579187a81cfd0b 2025-09-07T09:50:26.4699210Z deleted: sha256:5c6ef06b3536a430194aee509a784ee889c4a9d6248cb20fd9290e87e4ee2245 2025-09-07T09:50:26.4699591Z deleted: sha256:461aea034a25a2d72be6adfe9213c457c4cbf48724e9cb1c57987afb87668f21 2025-09-07T09:50:26.4699979Z deleted: sha256:e342cd1c71b7d0b024ea16b4a11f3f7fbbc2e3d11ef754c9d242aa50c4f8b0a3 2025-09-07T09:50:26.4700372Z deleted: sha256:bffd35a7fa1ddcfe05f79b7d3cae4180928eeea00eaab7ed7f484bc31adfc1d5 2025-09-07T09:50:26.4700803Z deleted: sha256:b34e33e7b04b5cbb5d5852199430593bfa18ddfe9081df42284230a14ebb739e 2025-09-07T09:50:26.4701174Z deleted: sha256:21d9b55338774d9ddc66d0bfcc92af9c8d2ecd94d1710b7049f5a811e411af7b 2025-09-07T09:50:26.4701558Z deleted: sha256:6cc2b33909585d17bf269fb8297ff881249e136137254734f7d23b9583208718 2025-09-07T09:50:26.4701940Z deleted: sha256:ca7f55b7c6d6cb11ddd8e187da34c2695fc2ce7655d652b9c9dc140a01ed056f 2025-09-07T09:50:26.4702334Z deleted: sha256:a3ece3d0ab6e99ef783c4f8d27d0e38504ab4477590ef556c16d22d92ba63a43 2025-09-07T09:50:26.4702712Z deleted: sha256:c137b0d41177c753aa1b69b11d0dd1f82420bf8520371866c845b53dca10b2d0 2025-09-07T09:50:26.4703356Z deleted: sha256:1e0d92b07bce12e511af59f608edd1932b10704d700f5e7538e406b90ecbb615 2025-09-07T09:50:26.4703737Z deleted: sha256:2ec3d01b3031e9da124d67410f54866ec5c679a0d6e4aee6b31608c45ce7fd77 2025-09-07T09:50:26.4704116Z deleted: sha256:308cffbd71363688c672b2043c6b9bf647cfb84593c42c3d88e3f36ee8f7f1b4 2025-09-07T09:50:26.4704496Z deleted: sha256:d965d9873fa450daba50a85d961f0835b14374167d84cfafa6060d16229f4229 2025-09-07T09:50:26.4704875Z deleted: sha256:effd997e222f62a34133bb2ecf9c0ffee151e5797f72e734d86a270d2e722374 2025-09-07T09:50:26.4705272Z deleted: sha256:0bbc1c78c10ee09c2697cfcce347dc9edbf82a7ccc25a6db6ee0a8dda398f7f2 2025-09-07T09:50:26.4705660Z deleted: sha256:214858e773d1ad73c2965c19b29cbfd3e2a974daa879163e1c1eb96567a7ee06 2025-09-07T09:50:26.4706041Z deleted: sha256:a9c7a2cd7ae229b26e84c093de657d0f4334d6cc9301991c6c3245ff62a9a71d 2025-09-07T09:50:26.4706424Z deleted: sha256:749a80551ef3f272e2517cb065bc7a5250da47d0b36bf74ed453caa9a5fee265 2025-09-07T09:50:26.4706797Z deleted: sha256:39b014c4e62d21c11df6c6d775d3f345675014292198981f455bacc4515a0f7b 2025-09-07T09:50:26.4707167Z deleted: sha256:0f087c9a894566644f825f5f87308d92e4cf149c51f7cd4769cbfaeefd3df791 2025-09-07T09:50:26.4707544Z deleted: sha256:dc6eb6dad5f9e332f00af553440e857b1467db1be43dd910cdb6830ba0898d50 2025-09-07T09:50:26.4707768Z 2025-09-07T09:50:26.4707863Z Total reclaimed space: 62.83GB 2025-09-07T09:50:26.4779885Z Post job cleanup. 2025-09-07T09:50:26.4809917Z Post job cleanup. 2025-09-07T09:50:26.5556194Z [command]/usr/bin/git version 2025-09-07T09:50:26.5598824Z git version 2.47.1 2025-09-07T09:50:26.5629317Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/9fca48f9-8840-46a9-a73f-63f2cb79e3ff/.gitconfig' 2025-09-07T09:50:26.5639041Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/9fca48f9-8840-46a9-a73f-63f2cb79e3ff' before making global git config changes 2025-09-07T09:50:26.5639686Z Adding repository directory to the temporary git global config as a safe directory 2025-09-07T09:50:26.5642526Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-09-07T09:50:26.5681081Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-09-07T09:50:26.5723733Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-09-07T09:50:26.6089649Z Entering 'android/libs/fbjni' 2025-09-07T09:50:26.6157569Z Entering 'third_party/FP16' 2025-09-07T09:50:26.6225977Z Entering 'third_party/FXdiv' 2025-09-07T09:50:26.6290153Z Entering 'third_party/NNPACK' 2025-09-07T09:50:26.6354728Z Entering 'third_party/NVTX' 2025-09-07T09:50:26.6422834Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T09:50:26.6496138Z Entering 'third_party/XNNPACK' 2025-09-07T09:50:26.6576124Z Entering 'third_party/aiter' 2025-09-07T09:50:26.6645267Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T09:50:26.6719063Z Entering 'third_party/benchmark' 2025-09-07T09:50:26.6785016Z Entering 'third_party/composable_kernel' 2025-09-07T09:50:26.6857644Z Entering 'third_party/cpp-httplib' 2025-09-07T09:50:26.6923414Z Entering 'third_party/cpuinfo' 2025-09-07T09:50:26.6987764Z Entering 'third_party/cudnn_frontend' 2025-09-07T09:50:26.7054185Z Entering 'third_party/cutlass' 2025-09-07T09:50:26.7133816Z Entering 'third_party/fbgemm' 2025-09-07T09:50:26.7203763Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T09:50:26.7267395Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T09:50:26.7339499Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T09:50:26.7402387Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T09:50:26.7472564Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T09:50:26.7540553Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T09:50:26.7602422Z Entering 'third_party/fbgemm/external/json' 2025-09-07T09:50:26.7670631Z Entering 'third_party/flash-attention' 2025-09-07T09:50:26.7738675Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T09:50:26.7808519Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T09:50:26.7882976Z Entering 'third_party/flatbuffers' 2025-09-07T09:50:26.7952316Z Entering 'third_party/fmt' 2025-09-07T09:50:26.8020558Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T09:50:26.8086006Z Entering 'third_party/gloo' 2025-09-07T09:50:26.8151791Z Entering 'third_party/googletest' 2025-09-07T09:50:26.8217813Z Entering 'third_party/ideep' 2025-09-07T09:50:26.8282257Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T09:50:26.8352466Z Entering 'third_party/ittapi' 2025-09-07T09:50:26.8418364Z Entering 'third_party/kineto' 2025-09-07T09:50:26.8485321Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T09:50:26.8548581Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T09:50:26.8613044Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T09:50:26.8675336Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T09:50:26.8740742Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T09:50:26.8810789Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T09:50:26.8879331Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T09:50:26.8943880Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T09:50:26.9013537Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T09:50:26.9078192Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T09:50:26.9145990Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T09:50:26.9209820Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T09:50:26.9275663Z Entering 'third_party/kleidiai' 2025-09-07T09:50:26.9343583Z Entering 'third_party/mimalloc' 2025-09-07T09:50:26.9408671Z Entering 'third_party/nlohmann' 2025-09-07T09:50:26.9475968Z Entering 'third_party/onnx' 2025-09-07T09:50:26.9557402Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T09:50:26.9626810Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T09:50:26.9693562Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T09:50:26.9759873Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T09:50:26.9824730Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T09:50:26.9888223Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T09:50:26.9955800Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T09:50:27.0019512Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T09:50:27.0082646Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T09:50:27.0145767Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T09:50:27.0209941Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T09:50:27.0276385Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T09:50:27.0357555Z Entering 'third_party/pocketfft' 2025-09-07T09:50:27.0424497Z Entering 'third_party/protobuf' 2025-09-07T09:50:27.0494596Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T09:50:27.0560215Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T09:50:27.0626822Z Entering 'third_party/psimd' 2025-09-07T09:50:27.0694704Z Entering 'third_party/pthreadpool' 2025-09-07T09:50:27.0764648Z Entering 'third_party/pybind11' 2025-09-07T09:50:27.0832751Z Entering 'third_party/python-peachpy' 2025-09-07T09:50:27.0900146Z Entering 'third_party/sleef' 2025-09-07T09:50:27.0965947Z Entering 'third_party/tensorpipe' 2025-09-07T09:50:27.1031941Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T09:50:27.1093656Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T09:50:27.1157584Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T09:50:27.1220169Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T09:50:27.1285354Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T09:50:27.1378353Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-09-07T09:50:27.1404166Z http.https://github.com/.extraheader 2025-09-07T09:50:27.1414607Z [command]/usr/bin/git config --local --unset-all http.https://github.com/.extraheader 2025-09-07T09:50:27.1447178Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-09-07T09:50:27.1799938Z Entering 'android/libs/fbjni' 2025-09-07T09:50:27.1848938Z http.https://github.com/.extraheader 2025-09-07T09:50:27.1889388Z Entering 'third_party/FP16' 2025-09-07T09:50:27.1937463Z http.https://github.com/.extraheader 2025-09-07T09:50:27.1979752Z Entering 'third_party/FXdiv' 2025-09-07T09:50:27.2024507Z http.https://github.com/.extraheader 2025-09-07T09:50:27.2064684Z Entering 'third_party/NNPACK' 2025-09-07T09:50:27.2111017Z http.https://github.com/.extraheader 2025-09-07T09:50:27.2149939Z Entering 'third_party/NVTX' 2025-09-07T09:50:27.2195021Z http.https://github.com/.extraheader 2025-09-07T09:50:27.2238037Z Entering 'third_party/VulkanMemoryAllocator' 2025-09-07T09:50:27.2281841Z http.https://github.com/.extraheader 2025-09-07T09:50:27.2324122Z Entering 'third_party/XNNPACK' 2025-09-07T09:50:27.2370605Z http.https://github.com/.extraheader 2025-09-07T09:50:27.2426768Z Entering 'third_party/aiter' 2025-09-07T09:50:27.2471056Z http.https://github.com/.extraheader 2025-09-07T09:50:27.2513187Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-09-07T09:50:27.2556645Z http.https://github.com/.extraheader 2025-09-07T09:50:27.2605281Z Entering 'third_party/benchmark' 2025-09-07T09:50:27.2648329Z http.https://github.com/.extraheader 2025-09-07T09:50:27.2688278Z Entering 'third_party/composable_kernel' 2025-09-07T09:50:27.2734381Z http.https://github.com/.extraheader 2025-09-07T09:50:27.2780254Z Entering 'third_party/cpp-httplib' 2025-09-07T09:50:27.2824665Z http.https://github.com/.extraheader 2025-09-07T09:50:27.2865597Z Entering 'third_party/cpuinfo' 2025-09-07T09:50:27.2911294Z http.https://github.com/.extraheader 2025-09-07T09:50:27.2951850Z Entering 'third_party/cudnn_frontend' 2025-09-07T09:50:27.2999242Z http.https://github.com/.extraheader 2025-09-07T09:50:27.3042276Z Entering 'third_party/cutlass' 2025-09-07T09:50:27.3086609Z http.https://github.com/.extraheader 2025-09-07T09:50:27.3134941Z Entering 'third_party/fbgemm' 2025-09-07T09:50:27.3180044Z http.https://github.com/.extraheader 2025-09-07T09:50:27.3226872Z Entering 'third_party/fbgemm/external/asmjit' 2025-09-07T09:50:27.3269184Z http.https://github.com/.extraheader 2025-09-07T09:50:27.3310077Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-09-07T09:50:27.3354386Z http.https://github.com/.extraheader 2025-09-07T09:50:27.3399924Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-09-07T09:50:27.3445268Z http.https://github.com/.extraheader 2025-09-07T09:50:27.3486349Z Entering 'third_party/fbgemm/external/cutlass' 2025-09-07T09:50:27.3532453Z http.https://github.com/.extraheader 2025-09-07T09:50:27.3582687Z Entering 'third_party/fbgemm/external/googletest' 2025-09-07T09:50:27.3625498Z http.https://github.com/.extraheader 2025-09-07T09:50:27.3665649Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-09-07T09:50:27.3712071Z http.https://github.com/.extraheader 2025-09-07T09:50:27.3750057Z Entering 'third_party/fbgemm/external/json' 2025-09-07T09:50:27.3794584Z http.https://github.com/.extraheader 2025-09-07T09:50:27.3840838Z Entering 'third_party/flash-attention' 2025-09-07T09:50:27.3886498Z http.https://github.com/.extraheader 2025-09-07T09:50:27.3931001Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-09-07T09:50:27.3973987Z http.https://github.com/.extraheader 2025-09-07T09:50:27.4020642Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-09-07T09:50:27.4063877Z http.https://github.com/.extraheader 2025-09-07T09:50:27.4112052Z Entering 'third_party/flatbuffers' 2025-09-07T09:50:27.4159482Z http.https://github.com/.extraheader 2025-09-07T09:50:27.4201887Z Entering 'third_party/fmt' 2025-09-07T09:50:27.4246526Z http.https://github.com/.extraheader 2025-09-07T09:50:27.4291446Z Entering 'third_party/gemmlowp/gemmlowp' 2025-09-07T09:50:27.4336215Z http.https://github.com/.extraheader 2025-09-07T09:50:27.4377993Z Entering 'third_party/gloo' 2025-09-07T09:50:27.4423780Z http.https://github.com/.extraheader 2025-09-07T09:50:27.4464406Z Entering 'third_party/googletest' 2025-09-07T09:50:27.4510526Z http.https://github.com/.extraheader 2025-09-07T09:50:27.4550207Z Entering 'third_party/ideep' 2025-09-07T09:50:27.4594099Z http.https://github.com/.extraheader 2025-09-07T09:50:27.4635359Z Entering 'third_party/ideep/mkl-dnn' 2025-09-07T09:50:27.4679727Z http.https://github.com/.extraheader 2025-09-07T09:50:27.4727099Z Entering 'third_party/ittapi' 2025-09-07T09:50:27.4771738Z http.https://github.com/.extraheader 2025-09-07T09:50:27.4812223Z Entering 'third_party/kineto' 2025-09-07T09:50:27.4857817Z http.https://github.com/.extraheader 2025-09-07T09:50:27.4898783Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-09-07T09:50:27.4942482Z http.https://github.com/.extraheader 2025-09-07T09:50:27.4983802Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-09-07T09:50:27.5027289Z http.https://github.com/.extraheader 2025-09-07T09:50:27.5069735Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-09-07T09:50:27.5116247Z http.https://github.com/.extraheader 2025-09-07T09:50:27.5156911Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-09-07T09:50:27.5206788Z http.https://github.com/.extraheader 2025-09-07T09:50:27.5246515Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-09-07T09:50:27.5290128Z http.https://github.com/.extraheader 2025-09-07T09:50:27.5330127Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-09-07T09:50:27.5371791Z http.https://github.com/.extraheader 2025-09-07T09:50:27.5417501Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-09-07T09:50:27.5461017Z http.https://github.com/.extraheader 2025-09-07T09:50:27.5502880Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-09-07T09:50:27.5547802Z http.https://github.com/.extraheader 2025-09-07T09:50:27.5588258Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-09-07T09:50:27.5635744Z http.https://github.com/.extraheader 2025-09-07T09:50:27.5677626Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-09-07T09:50:27.5720915Z http.https://github.com/.extraheader 2025-09-07T09:50:27.5766117Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-09-07T09:50:27.5811436Z http.https://github.com/.extraheader 2025-09-07T09:50:27.5850837Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-09-07T09:50:27.5893061Z http.https://github.com/.extraheader 2025-09-07T09:50:27.5989011Z Entering 'third_party/kleidiai' 2025-09-07T09:50:27.6034065Z http.https://github.com/.extraheader 2025-09-07T09:50:27.6077273Z Entering 'third_party/mimalloc' 2025-09-07T09:50:27.6120453Z http.https://github.com/.extraheader 2025-09-07T09:50:27.6161441Z Entering 'third_party/nlohmann' 2025-09-07T09:50:27.6207659Z http.https://github.com/.extraheader 2025-09-07T09:50:27.6251709Z Entering 'third_party/onnx' 2025-09-07T09:50:27.6296890Z http.https://github.com/.extraheader 2025-09-07T09:50:27.6353414Z Entering 'third_party/onnx/third_party/pybind11' 2025-09-07T09:50:27.6398170Z http.https://github.com/.extraheader 2025-09-07T09:50:27.6446985Z Entering 'third_party/opentelemetry-cpp' 2025-09-07T09:50:27.6491190Z http.https://github.com/.extraheader 2025-09-07T09:50:27.6535101Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-09-07T09:50:27.6576340Z http.https://github.com/.extraheader 2025-09-07T09:50:27.6618431Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-09-07T09:50:27.6660682Z http.https://github.com/.extraheader 2025-09-07T09:50:27.6702274Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-09-07T09:50:27.6746363Z http.https://github.com/.extraheader 2025-09-07T09:50:27.6786456Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-09-07T09:50:27.6831385Z http.https://github.com/.extraheader 2025-09-07T09:50:27.6873176Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-09-07T09:50:27.6917388Z http.https://github.com/.extraheader 2025-09-07T09:50:27.6955544Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-09-07T09:50:27.6998315Z http.https://github.com/.extraheader 2025-09-07T09:50:27.7039324Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-09-07T09:50:27.7081654Z http.https://github.com/.extraheader 2025-09-07T09:50:27.7121638Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-09-07T09:50:27.7162757Z http.https://github.com/.extraheader 2025-09-07T09:50:27.7208166Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-09-07T09:50:27.7252515Z http.https://github.com/.extraheader 2025-09-07T09:50:27.7294945Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-09-07T09:50:27.7340935Z http.https://github.com/.extraheader 2025-09-07T09:50:27.7399206Z Entering 'third_party/pocketfft' 2025-09-07T09:50:27.7444978Z http.https://github.com/.extraheader 2025-09-07T09:50:27.7487217Z Entering 'third_party/protobuf' 2025-09-07T09:50:27.7531776Z http.https://github.com/.extraheader 2025-09-07T09:50:27.7574407Z Entering 'third_party/protobuf/third_party/benchmark' 2025-09-07T09:50:27.7617607Z http.https://github.com/.extraheader 2025-09-07T09:50:27.7661576Z Entering 'third_party/protobuf/third_party/googletest' 2025-09-07T09:50:27.7706283Z http.https://github.com/.extraheader 2025-09-07T09:50:27.7747909Z Entering 'third_party/psimd' 2025-09-07T09:50:27.7790853Z http.https://github.com/.extraheader 2025-09-07T09:50:27.7835367Z Entering 'third_party/pthreadpool' 2025-09-07T09:50:27.7881236Z http.https://github.com/.extraheader 2025-09-07T09:50:27.7920800Z Entering 'third_party/pybind11' 2025-09-07T09:50:27.7965101Z http.https://github.com/.extraheader 2025-09-07T09:50:27.8007669Z Entering 'third_party/python-peachpy' 2025-09-07T09:50:27.8051365Z http.https://github.com/.extraheader 2025-09-07T09:50:27.8093525Z Entering 'third_party/sleef' 2025-09-07T09:50:27.8141244Z http.https://github.com/.extraheader 2025-09-07T09:50:27.8181277Z Entering 'third_party/tensorpipe' 2025-09-07T09:50:27.8228632Z http.https://github.com/.extraheader 2025-09-07T09:50:27.8269175Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-09-07T09:50:27.8311120Z http.https://github.com/.extraheader 2025-09-07T09:50:27.8352012Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-09-07T09:50:27.8396647Z http.https://github.com/.extraheader 2025-09-07T09:50:27.8438032Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-09-07T09:50:27.8482139Z http.https://github.com/.extraheader 2025-09-07T09:50:27.8523971Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-09-07T09:50:27.8567395Z http.https://github.com/.extraheader 2025-09-07T09:50:27.8606224Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-09-07T09:50:27.8648211Z http.https://github.com/.extraheader 2025-09-07T09:50:27.8787967Z A job completed hook has been configured by the self-hosted runner administrator 2025-09-07T09:50:27.8804616Z ##[group]Run '/home/ec2-user/runner-scripts/after_job.sh' 2025-09-07T09:50:27.8810436Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-09-07T09:50:27.8810682Z ##[endgroup] 2025-09-07T09:50:27.8910307Z [!ALERT!] Swap in detected! [!ALERT!] 2025-09-07T09:50:36.6314317Z [!ALERT!] Swap out detected [!ALERT!] 2025-09-07T09:50:51.8479712Z Cleaning up orphan processes